Announcing Timaeus
By Stan van Wingerden, Jesse Hoogland, Alexander Gietelink Oldenziel @ 2023-10-22T13:32 (+78)
This is a linkpost to https://www.lesswrong.com/posts/nN7bHuHZYaWv9RDJL/announcing-timaeus
Timaeus is a new AI safety research organization dedicated to making fundamental breakthroughs in technical AI alignment using deep ideas from mathematics and the sciences. Currently, we are working on singular learning theory and developmental interpretability. Over time we expect to work on a broader research agenda, and to create understanding-based evals informed by our research.
Activities
Our primary focus is research and research-training. For now, we're a remote-first organization. We collaborate primarily through online seminars and the DevInterp Discord, with regular in-person meetings at workshops and conferences (see below). We're also investing time in academic outreach to increase the general capacity for work in technical AI alignment.
Research
We believe singular learning theory, a mathematical subject founded by Sumio Watanabe, will lead to a better fundamental understanding of large-scale learning machines and the computational structures that they learn to represent. It has already given us concepts like the learning coefficient and insights into phase transitions in Bayesian learning. We expect significant advances in the theory to be possible, and that these advances can inform new tools for alignment.
Developmental interpretability is an approach to understanding the emergence of structure in neural networks, which is informed by singular learning theory but also draws on mechanistic interpretability and ideas from statistical physics and developmental biology. The key idea is that phase transitions organize learning and that detecting, locating, and understanding these transitions could pave a road to evaluation tools that prevent the development of dangerous capabilities, values, and behaviors. We're engaged in a research sprint to test the assumptions of this approach.
We see these as two particularly promising research directions, and they are our focus for now. Like any ambitious research, they are not guaranteed to succeed, but there's plenty more water in the well. Broadly speaking, the research agenda of Timaeus is oriented towards solving problems in technical AI alignment using deep ideas from across many areas of mathematics and the sciences, with a "full stack" approach that integrates work from pure mathematics through to machine learning experiments.
The outputs we have contributed to so far:
- Lau et al. (2023) "Quantifying Degeneracy in Singular Models via the Learning Coefficient"
- Chen et al. (2023) "Dynamical versus Bayesian Phase Transitions in a Toy Model of Superposition"
- Hoogland & Van Wingerden (2023), a distillation of Lau et al
- The DevInterp Repo & Python Package
- The 2023 Primer on Singular Learning Theory and Alignment
Academic Outreach
AI safety remains bottlenecked on senior researchers and mentorship capacity. The young people already in the field will grow into these roles. However, given the scale and urgency of the problem, we think it is important to open inroads to academia and encourage established scientists to spend their time on AI safety.
Singular learning theory and developmental interpretability can serve as a natural bridge between the emerging discipline of AI alignment and existing disciplines of mathematics and science, including physics and biology. We plan to spend part of our time onboarding scientists into alignment via concrete projects in these areas.
As part of these efforts, we're aiming to pilot a course at the University of Melbourne in late 2024 on "Mathematical Methods in AI Safety", similar in spirit to existing graduate courses on Mathematical Methods in Physics.
Conferences
We're organizing conferences, retreats, hackathons, etc. focusing on singular learning theory and developmental interpretability. These have included and will include:
- Conference on singular learning theory and AI alignment (Berkeley, June-July 2023)
- Retreat on singular learning theory (Amsterdam, September 2023)
- Hackathon on developmental interpretability (Melbourne, October 2023)
- Conference on developmental interpretability (Oxford, November 2023)
- AI safety demo days (Virtual, November 2023) open to independent and junior researchers to present their work.
- We've been amassing a list of project ideas to inspire researchers.
Team
Core Team
- Daniel Murfet (Research Director) is a mathematician at the University of Melbourne, an expert in singular learning theory, algebraic geometry, and mathematical logic.
- Jesse Hoogland (Executive Director) has a MSc. in theoretical physics from the University of Amsterdam and previously cofounded a health-tech startup before going all in on AI safety. He participated in MATS 3.0 & 3.1 under the supervision of Evan Hubinger and was a research assistant in David Krueger's lab.
- Stan van Wingerden (Operations & Finances) has a MSc. in theoretical physics from the University of Amsterdam and was previously the CTO of an algorithmic trading firm.
- Alexander Gietelink Oldenziel (Strategy & Outreach) is a DPhil student at University College London with a background in mathematics.
Research Assistants
We just concluded a round of hiring and are excited to bring on board several very talented Research Assistants (RAs), starting with
- Sai Niranjan
- George Wang
Friends and Collaborators
Since beginning to plan Timaeus in June 2023 we have been engaging with a range of people, both within the field of AI alignment and in academia. Here are some of the people we are actively collaborating with:
- Susan Wei (Statistician, University of Melbourne)
- Calin Lazaroiu (Mathematical Physicist, UNED-Madrid and IFIN-HH Bucharest)
- Simon Pepin Lehaulleur (Mathematician, Postdoc at University of Amsterdam)
- Tom Burns (Neuroscientist, Postdoc at ICERM, Brown University)
- Edmund Lau (PhD student, University of Melbourne)
- Zhongtian Chen (PhD student, University of Melbourne)
- Ben Gerraty (PhD student, University of Melbourne)
- Matthew Farrugia-Roberts (Research Assistant, David Krueger's AI Safety Lab, University of Cambridge)
- Liam Carroll (Independent AI safety researcher)
- Rohan Hitchcock (PhD student, University of Melbourne)
- Will Troiani (PhD student, University of Melbourne and University Sorbonne Paris Nord)
- Jake Mendel (Independent AI safety researcher)
- Zach Furman (Independent AI safety researcher)
Inclusion on this list does not imply endorsement of Timaeus' views.
Advisors
We're advised by Evan Hubinger and David ("Davidad") Dalrymple.
FAQ
Where can I learn more, and contact you?
Learn more on the Timaeus webpage. You can email Jesse Hoogland.
What about capabilities risk?
There is a risk that fundamental progress in either singular learning theory or developmental interpretability could contribute to further acceleration in AI capabilities in the medium term. We take this seriously and are seeking advice from other alignment researchers and organizations. By the end of our current research sprint we will have in place institutional forms to help us navigate this risk.
Likewise, there is a risk that outreach which aims to involve more scientists in AI alignment work will also accelerate progress in AI capabilities. However, those of us in academia can already see that as the risks become more visible, scientists are starting to think about these problems on their own. So the question is not whether a broad range of scientists will become interested in alignment but when they will start to contribute and what they work on.
It is part of Timaeus' mission to help scientists to responsibly contribute to technical AI alignment, while minimizing these risks.
Are phase transitions really the key?
The strongest critique of developmental interpretability we know is the following: while it is established that phase transitions exist in neural network training, it is not yet clear how common they are, and whether they make a good target for alignment.
We think developmental interpretability is a good investment in a world where many of the important structures (e.g., circuits) in neural networks form in phase transitions. Figuring out whether we live in such a world is one of our top priorities. It's not trivial because even if transitions exist they may not necessarily be visible to naive probes. Our approach is to systematically advance the fundamental science of finding and classifying transitions, starting with smaller systems where transitions can be definitively shown to exist.
How are you funded?
We're funded through a $142k Manifund grant led primarily by Evan Hubinger, Ryan Kidd, Rachel Weinberg, and Marcus Abramovitch. We are fiscally sponsored by Ashgro. We could put an additional $500k to work. (If you're interested in contributing, reach out to Jesse Hoogland.) If our research pans out as we anticipate, we'll be aiming to raise $1-5m in Q2 of 2024.
"Timaeus"? How do I even pronounce that?
Pronounce it however you want.
Timaeus is the eponymous character in the dialogue where Plato introduces his theory of forms. The dialogue posits a correspondence between the elements that make up the world and the Platonic solids. That's wrong, but it contains the germ of the idea of the unreasonable effectiveness of mathematics in understanding the natural world.
We read the Timaeus dialogue with a spirit of hope, in the capacity of the human intellect to understand and solve wicked problems. The narrow gate to human flourishing is preceded by a narrow path.
We'll see you on that path.