Papers
Topics
Authors
Recent
Search
2000 character limit reached

Dynamical stability and chaos in artificial neural network trajectories along training

Published 8 Apr 2024 in cs.LG, cond-mat.dis-nn, nlin.CD, and physics.data-an | (2404.05782v1)

Abstract: The process of training an artificial neural network involves iteratively adapting its parameters so as to minimize the error of the network's prediction, when confronted with a learning task. This iterative change can be naturally interpreted as a trajectory in network space -- a time series of networks -- and thus the training algorithm (e.g. gradient descent optimization of a suitable loss function) can be interpreted as a dynamical system in graph space. In order to illustrate this interpretation, here we study the dynamical properties of this process by analyzing through this lens the network trajectories of a shallow neural network, and its evolution through learning a simple classification task. We systematically consider different ranges of the learning rate and explore both the dynamical and orbital stability of the resulting network trajectories, finding hints of regular and chaotic behavior depending on the learning rate regime. Our findings are put in contrast to common wisdom on convergence properties of neural networks and dynamical systems theory. This work also contributes to the cross-fertilization of ideas between dynamical systems theory, network theory and machine learning

Definition Search Book Streamline Icon: https://streamlinehq.com
References (57)
  1. Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT press, Cambridge (2016) Marcus [2018] Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  2. Marcus, G.: Deep Learning: A Critical Appraisal. arXiv (2018) https://doi.org/10.48550/ARXIV.1801.00631 San Miguel [2023] San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  3. San Miguel, M.: Frontiers in complex systems. Frontiers in Complex Systems 1, 1080801 (2023) Arola-Fernández and Lacasa [2023] Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  4. Arola-Fernández, L., Lacasa, L.: An effective theory of collective deep learning. arXiv preprint arXiv:2310.12802 (2023) Prisner [1995] Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  5. Prisner, E.: Graph Dynamics (Pitman Research Notes in Mathematics Series). Chapman & Hall CRC, London (1995) Ruder [2016] Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  6. Ruder, S.: An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016) Hoffer et al. [2017] Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  7. Hoffer, E., Hubara, I., Soudry, D.: Train longer, generalize better: closing the generalization gap in large batch training of neural networks. Advances in neural information processing systems 30 (2017) Holme and Saramäki [2019] Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  8. Holme, P., Saramäki, J.: Temporal Network Theory vol. 2. Springer, New York (2019) Lacasa et al. [2022] Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  9. Lacasa, L., Rodriguez, J.P., Eguiluz, V.M.: Correlations of network trajectories. Physical Review Research 4(4), 042008 (2022) https://doi.org/10.1103/PhysRevResearch.4.L042008 Caligiuri et al. [2023] Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  10. Caligiuri, A., Eguíluz, V.M., Di Gaetano, L., Galla, T., Lacasa, L.: Lyapunov exponents for temporal networks. Physical Review E 107(4), 044305 (2023) https://doi.org/10.1103/PhysRevE.107.044305 La Malfa et al. [2022] La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  11. La Malfa, E., La Malfa, G., Caprioli, C., Nicosia, G., Latora, V.: Deep Neural Networks as Complex Networks. arxiv (2022) https://doi.org/10.48550/ARXIV.2209.05488 La Malfa et al. [2021] La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  12. La Malfa, E., La Malfa, G., Nicosia, G., Latora, V.: Characterizing Learning Dynamics of Deep Neural Networks via Complex Networks. In: 2021 IEEE 33rd International Conference on Tools with Artificial Intelligence (ICTAI), pp. 344–351. IEEE, Washington, DC, USA (2021). https://doi.org/10.1109/ICTAI52525.2021.00056 . https://ieeexplore.ieee.org/document/9643155/ Ribas et al. [2020] Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  13. Ribas, L.C., Sá Junior, J.J.D.M., Scabini, L.F.S., Bruno, O.M.: Fusion of complex networks and randomized neural networks for texture analysis. Pattern Recognition 103, 107189 (2020) https://doi.org/10.1016/j.patcog.2019.107189 Scabini et al. [2022] Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  14. Scabini, L., De Baets, B., Bruno, O.M.: Improving Deep Neural Network Random Initialization Through Neuronal Rewiring. arxiv (2022) https://doi.org/10.48550/ARXIV.2207.08148 Schuster and Just [2006] Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  15. Schuster, H.G., Just, W.: Deterministic Chaos: an Introduction. John Wiley & Sons, Weinheim (2006) Bak et al. [1988] Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  16. Bak, P., Tang, C., Wiesenfeld, K.: Self-organized criticality. Physical review A 38(1), 364 (1988) Langton [1990] Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  17. Langton, C.G.: Computation at the edge of chaos: Phase transitions and emergent computation. Physica D: nonlinear phenomena 42(1-3), 12–37 (1990) Carroll [2020] Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  18. Carroll, T.L.: Do reservoir computers work best at the edge of chaos? Chaos: An Interdisciplinary Journal of Nonlinear Science 30(12) (2020) Fisher [1936] Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  19. Fisher, R.A.: The use of multiple measurements in taxonomic problems. Annals of eugenics 7(2), 179–188 (1936) Ziyin et al. [2023] Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  20. Ziyin, L., Li, B., Galanti, T., Ueda, M.: The probabilistic stability of stochastic gradient descent. arXiv preprint arXiv:2303.13093 (2023) Strogatz [2015] Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  21. Strogatz, S.H.: Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering, Second edition edn. Westview Press, a member of the Perseus Books Group, Boulder, CO (2015) Aurell et al. [1996] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  22. Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Growth of noninfinitesimal perturbations in turbulence. Physical review letters 77(7), 1262 (1996) Aurell et al. [1997] Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  23. Aurell, E., Boffetta, G., Crisanti, A., Paladin, G., Vulpiani, A.: Predictability in the large: an extension of the concept of lyapunov exponent. Journal of physics A: Mathematical and general 30(1), 1 (1997) Montana et al. [1989] Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  24. Montana, D.J., Davis, L., et al.: Training feedforward neural networks using genetic algorithms. In: IJCAI, vol. 89, pp. 762–767 (1989) Choromanska et al. [2015] Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  25. Choromanska, A., Henaff, M., Mathieu, M., Ben Arous, G., LeCun, Y.: The loss surfaces of multilayer networks. In: Lebanon, G., Vishwanathan, S.V.N. (eds.) Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics. Proceedings of Machine Learning Research, vol. 38, pp. 192–204. PMLR, San Diego, California, USA (2015). https://proceedings.mlr.press/v38/choromanska15.html Cohen et al. [2021] Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  26. Cohen, J., Kaur, S., Li, Y., Kolter, J.Z., Talwalkar, A.: Gradient descent on neural networks typically occurs at the edge of stability. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=jh-rTtvkGeM Kong and Tao [2020] Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  27. Kong, L., Tao, M.: Stochasticity of deterministic gradient descent: Large learning rate for multiscale objective function. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H. (eds.) Advances in Neural Information Processing Systems, vol. 33, pp. 2625–2638. Curran Associates, Inc., New York, United States (2020). https://proceedings.neurips.cc/paper_files/paper/2020/file/1b9a80606d74d3da6db2f1274557e644-Paper.pdf Agarwal et al. [2021] Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  28. Agarwal, N., Goel, S., Zhang, C.: Acceleration via fractal learning rate schedules. arxiv (2021) 2103.01338 Lee et al. [2016] Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  29. Lee, J.D., Simchowitz, M., Jordan, M.I., Recht, B.: Gradient Descent Only Converges to Minimizers. In: Conference on Learning Theory, pp. 1246–1257. PMLR, Colorado, USA (2016). https://proceedings.mlr.press/v49/lee16.html Kawaguchi [2016] Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  30. Kawaguchi, K.: Deep learning without poor local minima. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc., New York, United States (2016). https://proceedings.neurips.cc/paper_files/paper/2016/file/f2fc990265c712c49d51a18a32b39f0c-Paper.pdf Zhu et al. [2020] Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  31. Zhu, Z., Soudry, D., Eldar, Y.C., Wakin, M.B.: The Global Optimization Geometry of Shallow Linear Neural Networks. Journal of Mathematical Imaging and Vision 62(3), 279–292 (2020) https://doi.org/10.1007/s10851-019-00889-w Bosman et al. [2020] Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  32. Bosman, A.S., Engelbrecht, A., Helbig, M.: Visualising basins of attraction for the cross-entropy and the squared error neural network loss functions. Neurocomputing 400, 113–136 (2020) https://doi.org/10.1016/j.neucom.2020.02.113 Fort and Jastrzebski [2019] Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  33. Fort, S., Jastrzebski, S.: Large Scale Structure of Neural Network Loss Landscapes. arXiv (2019) https://doi.org/10.48550/ARXIV.1906.04724 Fort et al. [2019] Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  34. Fort, S., Hu, H., Lakshminarayanan, B.: Deep Ensembles: A Loss Landscape Perspective. arxiv (2019) https://doi.org/10.48550/ARXIV.1912.02757 Havasi et al. [2020] Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  35. Havasi, M., Jenatton, R., Fort, S., Liu, J.Z., Snoek, J., Lakshminarayanan, B., Dai, A.M., Tran, D.: Training independent subnetworks for robust prediction. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.06610 Fort et al. [2020] Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  36. Fort, S., Dziugaite, G.K., Paul, M., Kharaghani, S., Roy, D.M., Ganguli, S.: Deep learning versus kernel learning: An empirical study of loss landscape geometry and the time evolution of the Neural Tangent Kernel. arxiv (2020) https://doi.org/10.48550/ARXIV.2010.15110 Alligood et al. [1996] Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  37. Alligood, K.T., Sauer, T., Yorke, J.A.: Chaos: An Introduction to Dynamical Systems. Textbooks in Mathematical Sciences. Springer, New York (1996) Holmes and Shea-Brown [2006] Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  38. Holmes, P., Shea-Brown, E.T.: Stability. Scholarpedia 1(10), 1838 (2006) https://doi.org/10.4249/scholarpedia.1838 . revision #137538 Fyodorov and Khoruzhenko [2016] Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  39. Fyodorov, Y.V., Khoruzhenko, B.A.: Nonlinear analogue of the may-wigner instability transition. Proceedings of the National Academy of Sciences 113(25), 6827–6832 (2016) https://doi.org/10.1073/pnas.1601136113 Ben Arous et al. [2021] Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  40. Ben Arous, G., Fyodorov, Y.V., Khoruzhenko, B.A.: Counting equilibria of large complex systems by instability index. Proceedings of the National Academy of Sciences 118(34), 2023719118 (2021) https://doi.org/10.1073/pnas.2023719118 Bosman et al. [2020] Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  41. Bosman, A.S., Petrus Engelbrecht, A., Helbig, M.: Loss Surface Modality of Feed-Forward Neural Network Architectures. In: 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE, Glasgow, United Kingdom (2020). https://doi.org/10.1109/IJCNN48605.2020.9206727 . https://ieeexplore.ieee.org/document/9206727/ Núnez et al. [2013] Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  42. Núnez, A.M., Luque, B., Lacasa, L., Gómez, J.P., Robledo, A.: Horizontal visibility graphs generated by type-i intermittency. Physical Review E 87(5), 052801 (2013) Nunez et al. [2012] Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  43. Nunez, A., Lacasa, L., Valero, E., Gómez, J.P., Luque, B.: Detecting series periodicity with horizontal visibility graphs. International Journal of Bifurcation and Chaos 22(07), 1250160 (2012) Kantz [1994] Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  44. Kantz, H.: A robust method to estimate the maximal lyapunov exponent of a time series. Physics letters A 185(1), 77–87 (1994) [46] Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  45. Baldassi, C., Pittorino, F., Zecchina, R.: Shaping the learning landscape in neural networks around wide flat minima 117(1), 161–170 https://doi.org/10.1073/pnas.1908636117 . Accessed 2024-03-25 [47] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  46. Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition 86(11), 2278–2324 https://doi.org/10.1109/5.726791 . Accessed 2024-03-25 Boedecker et al. [2012] Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  47. Boedecker, J., Obst, O., Lizier, J.T., Mayer, N.M., Asada, M.: Information processing in echo state networks at the edge of chaos. Theory in Biosciences 131, 205–213 (2012) Vettelschoss et al. [2021] Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  48. Vettelschoss, B., Röhm, A., Soriano, M.C.: Information processing capacity of a single-node reservoir computer: An experimental evaluation. IEEE Transactions on Neural Networks and Learning Systems 33(6), 2714–2725 (2021) Watkins et al. [2016] Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  49. Watkins, N.W., Pruessner, G., Chapman, S.C., Crosby, N.B., Jensen, H.J.: 25 years of self-organized criticality: concepts and controversies. Space Science Reviews 198, 3–44 (2016) Hidalgo et al. [2014] Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  50. Hidalgo, J., Grilli, J., Suweis, S., Munoz, M.A., Banavar, J.R., Maritan, A.: Information-based fitness and the emergence of criticality in living systems. Proceedings of the National Academy of Sciences 111(28), 10095–10100 (2014) Munoz [2018] Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  51. Munoz, M.A.: Colloquium: Criticality and dynamical scaling in living systems. Reviews of Modern Physics 90(3), 031001 (2018) Chialvo [2010] Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  52. Chialvo, D.R.: Emergent complex neural dynamics. Nature physics 6(10), 744–750 (2010) Moretti and Muñoz [2013] Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  53. Moretti, P., Muñoz, M.A.: Griffiths phases and the stretching of criticality in brain networks. Nature communications 4(1), 2521 (2013) Morales et al. [2023a] Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  54. Morales, G.B., Di Santo, S., Muñoz, M.A.: Quasiuniversal scaling in mouse-brain neuronal activity stems from edge-of-instability critical dynamics. Proceedings of the National Academy of Sciences 120(9), 2208998120 (2023) Morales et al. [2023b] Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  55. Morales, G.B., Di Santo, S., Muñoz, M.A.: Unveiling the intrinsic dynamics of biological and artificial neural networks: from criticality to optimal representations. arXiv preprint arXiv:2307.10669 (2023) Morales and Muñoz [2021] Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  56. Morales, G.B., Muñoz, M.A.: Optimal input representation in neural systems at the edge of chaos. Biology 10(8), 702 (2021) Geiping et al. [2022] Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
  57. Geiping, J., Goldblum, M., Pope, P., Moeller, M., Goldstein, T.: Stochastic training is not necessary for generalization. In: International Conference on Learning Representations (2022). https://openreview.net/forum?id=ZBESeIUB5k
Citations (3)

Summary

  • The paper demonstrates that ANN training dynamics can alternate between convergence and divergence, even under low learning rate conditions.
  • It reveals that higher learning rates induce an 'edge of stability' behavior marked by fast but non-monotonic loss convergence and sensitive dependence on initial conditions.
  • Experiments with very high learning rates show quasi-periodic and chaotic dynamics, highlighting the need to optimize learning rate settings for robust performance.

Dynamical Stability and Chaos in ANN Trajectories During Training

This essay discusses the theoretical exploration of artificial neural network (ANN) training dynamics through the lens of dynamical systems theory. The paper attempts to deepen the understanding of complex behaviors exhibited during the ANN training process by analysing the evolution of network trajectories under different learning rate regimes. Here, a shallow neural network's behavior is scrutinized as it attempts to learn a simple classification task.

Introduction to ANN Dynamics

In ANN training, parameters are iteratively updated to minimize prediction error, construed as trajectories in graph space. By using dynamical systems theory, the research examines ANN weight changes as a time series, interpreting gradient descent optimization as a discrete dynamical system. Broadly, the paper investigates how these network trajectories behave and evolve, particularly focusing on dynamical and orbital stability concepts, which are crucial for understanding convergence properties in machine learning. Figure 1

Figure 1: The training process of an ANN is depicted as a network trajectory in graph space, where in each iteration of the optimization scheme the network parameters are updated, leading to a decreasing loss function.

Preliminaries

This section lays the groundwork for interpreting ANN training as a dynamical process. A one-hidden-layer, feed-forward neural network is trained with the Iris dataset, where the input comprises physiological properties of iris flowers and the task is species classification. The simple architecture with a single hidden layer and cross-entropy loss is used to elucidate the fundamental dynamics of network training. The gradient descent algorithm is employed with both small and large constant learning rates to observe convergence behavior. Figure 2

Figure 2: Illustration of the Iris dataset and difficulty in linearly separating the three classes. Datapoints are shown in the space of two of their four input features, namely "sepal length" and "sepal width". Colors correspond to different classes, while markers show whether the instances were classified correctly or not (marked as ‘x’ if the prediction was incorrect).

Low Learning Rate Regime

At a low learning rate (e.g., η=0.01\eta=0.01), gradient descent is expected to converge monotonically. However, this research reveals contrasting behavior: network trajectories do not converge smoothly to fixed points; instead, they alternate between convergence and divergence. Results imply that this could be due to flat loss landscapes or marginal stability where trajectories drift in parameter space without significant improvement in loss. Figure 3

Figure 3: Example showing the evolution of the distances between reference and perturbed trajectories, for a perturbation radius ϵ=10−8\epsilon=10^{-8}.

High Learning Rate Regime

Edge of Stability (η=1\eta=1)

Increasing the learning rate to η=1\eta=1 leads to a regime termed the 'edge of stability'. Here, the ANN does not diverge but rather converges faster despite a non-monotonic loss trajectory. A clear signature of sensitive dependence on initial conditions is observed, evidenced through the estimation of network Lyapunov exponents. Figure 4

Figure 4: Lyapunov exponents for trajectories in the Edge of Stability (eta=1) regime. (Left panel) Distribution of finite Lyapunov exponents P(\Lambda), where each Lambda is estimated from Eq. ${\ref{eq:Lambda}}$

Very Large Learning Rate (η=5\eta=5)

In this regime, the dynamics become significantly unpredictable. The complex behavior noted includes quasi-periodic and intermittent-like chaotic dynamics. These observations suggest a transition to chaotic behavior, with fascinating implications for understanding the dynamics of very high learning rate scenarios. Figure 5

Figure 5: Training trajectories in the very large eta=5 regime. Each column a)-d) represents the trajectory starting from an independent initial condition.

Discussion

The explorations presented demonstrate that ANN training dynamics can exhibit a complex array of behaviors, challenging simple convergence narratives. The phenomena observed at different learning rates underscore the theory's capacity to unravel underlying chaotic or complex dynamics, reflected in both marginal stability zones and the 'edge of stability'.

The findings reiterate the need to consider such dynamical behaviors when configuring learning rates and understanding potential ANN learning pathways, potentially optimizing search algorithms for machine learning applications.

Conclusion

The studied trajectories indicate that classical dynamical systems tools are powerful for interpreting ANN training processes, revealing consistent patterns across simple and complex learning tasks. Future work should aim at further intersecting the fields of machine learning, dynamical systems, and temporal networks to offer deeper insights and more effective learning strategies.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 9 tweets with 170 likes about this paper.