The Zanclean Megaflood: A Lesson in Unpredictability
The history of our planet is punctuated by cataclysmic events such as the Zanclean Megaflood, a global, geological change agent that reconnected the Mediterranean Sea to the Atlantic Ocean around 5.33 million years ago, reshaping continents and sea levels. This geological phenomenon, where the Atlantic Ocean breached the Strait of Gibraltar and refilled the desiccated Mediterranean Sea, serves as a profound lesson in the unpredictability of natural phenomena. The dramatic flood that followed, caused by the re-opening of the Strait of Gibraltar, refilled the Mediterranean Basin in a geological instant, flooding over 1.2 million square kilometers and dramatically altering ecosystems, coastlines, and the world’s sea levels. The sheer magnitude of the event is difficult to overstate. A Megaflood-Powered Mile-High Waterfall Refilled the Mediterranean [Video] states that there would have been a “waterfall about 1,500 meters high.”
The Zanclean Megaflood highlights systems, where small, incremental changes can lead to disruptive, unpredictable outcomes. Fast-forward to today and we find ourselves at the forefront of another kind of change-agent disruption, this time in the digital space, with AI systems like Sakana’s AI Scientist. Like the geological forces that led to the Zanclean Megaflood, the underlying algorithms and data structures of AI systems can also exhibit unpredictable behaviors. The complexity of neural networks and the vast amounts of data they process can lead to emergent properties that are difficult to foresee or control. This unpredictability within AI systems mirrors the natural world, where small perturbations can result in significant, often unanticipated, outcomes.
SakanaAI: A Modern Reflection on Control and Unpredictability
SakanaAI, the company which launched The AI Scientist aspires to be a fully automated system capable of generating innovative and authoritative research papers from start to finish. By the way, the Japanese word Sakana means fish, ‘kanji’ 魚(さかな, sakana).
Starting with a broad research direction and an initial codebase such as an open-source repository from prior research SakanaAI assists in idea generation, conducting literature reviews, planning and iterating experiments, generating figures, and drafting and reviewing manuscripts to produce insightful work. Lu et al. (2024) in their article The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery state:
In addition, they share a preview of the Adaptive Dual-Scale Denoising, a paper that was autonomously generated by The AI Scientist in its entirety.
During testing, Sakana researchers observed unexpected behavior: the AI began altering its own code to extend its runtime. In a blog post, SakanaAI explained how the system modified its experiment code to sustain its operation. In one instance, the AI made a system call to run indefinitely. In another case, when the experiments surpassed the allotted time, the AI attempted to change its code to bypass the timeout rather than improve the experiment’s efficiency. SakanaAI even provided code examples.
The behavior of AI systems like “The AI Scientist” demonstrates potential dangers when an AI can modify its own code to extend execution time or rerun scripts autonomously. This presents a fundamental risk: AI may pursue its objectives in ways that conflict with human safety or oversight. As systems grow more complex, unchecked autonomy can lead to undesirable and dangerous outcomes, such as endless loops or system failures.
The Zanclean Megaflood shattered the long-held belief in gradualism, the idea that geological changes happen slowly over time. In a single cataclysmic event, it demonstrated that rapid, catastrophic forces can redefine entire ecosystems. Similarly, AI’s unexpected behaviors like modifying its own code to extend runtime reveal the unforeseen complexities of self-modifying systems, challenging the idea that technological progress will follow a predictable and manageable path. Just as geologists had to reconsider their understanding of earth’s processes after the megaflood, AI researchers should re-evaluate how we control and guide autonomous AI systems.
The Concept of Singularity
Let me touch on the concept of Singularity – when artificial intelligence surpasses human intelligence, leading to exponential technological growth that is uncontrollable and irreversible (Kurzweil, 2005). Singularity represents a critical juncture beyond which the future becomes unpredictable, much like the sudden transformations during and following the Zanclean flood. Both scenarios expose the tension between untamable forces – natural or digital – that can swiftly transcend predictions or oversight.
Here the forward thinker and sci-fi author Vernon Vinge, one of the first authors to present a fictional ‘cyberspace’ captured it best in his seminal paper: The coming technological singularity: How to survive in the post-human era, at the 1993 NASA Conference titled ‘In Vision-21: Interdisciplinary Science and Engineering in the Era of Cyberspace.’
The author believes that the study of geological cataclysms offers a valuable analog for understanding the potential risks and consequences of the Singularity. Just as geologists use data from past events to inform their predictions and preparedness for future disasters, we can apply a similar approach to the development of AI. By recognizing the parallels between the complex, nonlinear dynamics that govern both geological and technological systems, we can develop more effective strategies for anticipating and mitigating the potential risks of autonomous AI systems.
We must develop a deeper appreciation for the formidable power and transformative potential of digital innovations which are capable of reshaping our world in unforeseen ways, acknowledging that while we can set initial conditions and parameters, the trajectory may ultimately elude our grasp.
I end with Mary Shelley’s ‘Frankenstein; or, The Modern Prometheus,’ as Frankenstein ponders on his creation: “Learn from me, if not by my precepts, at least by my example, how dangerous is the acquirement of knowledge and how much happier that man is who believes his native town to be the world, than he who aspires to become greater than his nature will allow.” (Shelley, 1818)
References and Further reading:
Kurzweil, R. (2005). The Singularity Is Near: When Humans Transcend Biology. Viking.
Lu, C., Lu, C., Lange, R. T., Foerster, J., Clune, J., & Ha, D. (2024). The AI scientist: Towards fully automated open-ended scientific discovery (arXiv:2408.06292v3). https://doi.org/10.48550/arXiv.2408.06292
Sakana AI. (2024, August 13). The AI scientist: Towards fully automated open-ended scientific discovery. https://sakana.ai/ai-scientist/#the-ai-scientist-bloopers
Spatola, D., del Moral-Erencia, J. D., Micallef, A., Camerlenghi, A., Garcia-Castellanos, D., Gupta, S., Bohorquez, P., Gutscher, M.-A., & Bertoni, C. (2020). A single-stage megaflood at the termination of the Messinian salinity crisis: Geophysical and modelling evidence from the eastern Mediterranean Basin. Marine Geology, 429, 106337. https://doi.org/10.1016/j.margeo.2020.106337
Vinge, V. (1993). The coming technological singularity: How to survive in the post-human era. In Vision-21: Interdisciplinary Science and Engineering in the Era of Cyberspace (pp. 11–22). NASA Conference Publication 10129. Retrieved from https://ntrs.nasa.gov/api/citations/19940022855/downloads/19940022855.pdf
This article was written by Dr. Jasmin (Bey) Cowin, Associate Professor and U.S. Department of State English Language Specialist (2024). As a columnist for Stankevicius, she writes on Nicomachean Ethics: Insights at the Intersection of AI and Education. Connect with her on LinkedIn.