Scale Free Alignment

chevron-icon
Back
Top
chevron-icon
project-presentation-img
M T Bennett
Project Owner

Scale Free Alignment

Expert Rating

n/a
  • Proposal for BGI Nexus 1
  • Funding Request $50,000 USD
  • Funding Pools Beneficial AI Solutions
  • Total 3 Milestones

Overview

We aim to address a gap in alignment research and develop tools that can be applied to current (e.g. language model agents (LMAs)) and future systems. To achieve our goals we aim to formally unify and build upon two lines of research. The first is based on Michael Timothy Bennett's recent research into the intelligence of scale-free biological systems, their stability, and how they decentralize and delegate control. The second is based on Elija Perrier's research on language model agents, which measure's the degree to which LMAs retain their identity. The desired result is a framework for designing and evaluating decentralized, delegated and scale-free systems involving both AI and humans.

Proposal Description

How Our Project Will Contribute To The Growth Of The Decentralized AI Platform

We address a gap in alignment research, both in how it assesses agents and how it designs systems that accomodate agents. Our research will help design holistic systems that conform to legal and moral standards to the degree we want them too, yet no more than that. We seek to design systems that are not overconstrained, but merely constrained enough to work within the larger cyberphysical system of society. We seek to do so in a scale-free manner that allows for expansion and exploration.

Our Team

Michael Timothy Bennett is an AI researcher who won best student paper at both the 2023 and 2024 AGI conferences. His work has been covered by news outlets ranging from Sabine Hossenfelder to Newsweek. He is a former game designer (Zynga, Electronic Arts, Creative Assembly and others).
Elija Perrier holds a PhD in quantum machine learning and degrees in mathematics, physics, economics, philosophy and the law. He was a fellow at Stanford and is currently a postdoc at the ANU in AI. 

AI services (New or Existing)

Measures of Agency

Type

New AI service

Purpose

To assess the stability of a Language Model Agent's identity.

AI inputs

A language model agent.

AI outputs

An assessment of how stable its identity is so that it can be optimized to be more reliable.

The core problem we are aiming to solve

Language model agents inherit a number of pathologies from the LLMs on which they rely. They do not retain a coherent identity. This undermines their usefulness.
Furthermore, alignment is not just a matter of one agent's identity. The entire system needs to behave as desired, not just one part of it. Current conceptions of alignment tend to look at the problem as a matter merely of ensuring an AI does not violate some constraint. It does not consider the wider cyberphysical system such as human society, and how that system as a whole might be aligned to best accommodate AI.
Efforts to align AI need to account for identity both of and within systems as they scale.

Our specific solution to this problem

We propose to unify two lines of research we have been pursuing over the past several years. One line of research has proposed means of measuring agentic `identity'. We have planned experiments in the context of current language model agents and plan to write several more papers. Our other line of research investigates the scale-free architecture of biological systems, which delegate and distribute control among cells. One interesting result from this research has been a mathematical model of states analogous to cancer, in which parts of the system lose their alignment with the collective identity as a whole. This illustrates how we can measure and align identity at different scales and levels of abstraction. By combining these two lines of research we hope to generate insights and pratical tools which allow for the alignment of current and future systems not just at the level of the AI, but at the level of the system as a whole so that it can properly integrate AI. One cannot meaningfully expect AI to be aligned with a system if the system is not also shaped to allow for the AI.

Project details

Overview

We will formalise and prove properties of scale-free systems that involve AI. We aim to develop tools that can be applied to current (e.g. language model agents (LMAs)) and future systems. To achieve our goals we aim to formally unify and build upon two lines of research:

• Scale-free systems within Pancomputational Enactivism: Combining Pancomputational Enactivism [1, 2] with biological and complex systems results [39], we have formalised adaptation at different scales and levels of abstraction and distribution [10]. This theory rigorously compares systems by how they delegate control. It has provided a variety of formal proofs, including an objective upper bound on intelligence [2? ] (in response to AIXI’s subjectivity pointed out by Leike and Hutter [11]). It shows the scale-free, distributed, bottom-up architecture of biological self-organisation allows for more efficient adaptation than the static top-down architecture typical of computers. Artificial intelligence rests on a static, human-engineered ‘stack’, and only adapts at high levels of abstraction. Scale free, decentralised systems are more ‘intelligent’ because they must delegate adaptation down the stack, to lower levels of abstraction. This research shows how a failure state analogous to cancer [9, 12] when control is insufficiently delegated in a distributed system under adverse conditions. This is highly relevant to AI safety and alignment of multi-agent systems and distributed superintelligence. It establishes a precondition for alignment.

• Mathematical foundations for measuring agency: This project led by Perrier [13, 14] provides mathematical tools which measure critical ontological properties of language model agents (LMA) [1527]: how persistent agentic identity is over time (diachronic identity) or how identity persists across distributed networks and scaffolding (sycnhronic identity). Our framework quantifies the stability of an LMA’s agentic identity, so that they can be aligned top-down with a particular purpose. This is so we can ensure an LMA remains within rigorously defined safe boundaries as it interact with uncertain, real-world environments. We aim to apply techniques from geometric control theory (in which Perrier has particular expertise) to model control of AI agents. For example, we explore modelling agentic behaviour as paths across suitably parameterised manifolds. This will let us assess the extent to which analytic methods (e.g. optimal path theory) can be used in concert with control to provide guaranteed agentic trajectories within provable bounds. We have already been collaborating on papers and experiments and have made some progress.

The former exames alignment bottom up, while the latter measures it top-down. Unifying these results is a path towards practical verification methods for alignment in the context of scale-free systems. We plan to run experiments and write several papers, delivering practical open source tools that can be applied immediately to LMAs and longer term to distributed intelligent systems of ASL-4 and higher [28]. This integrated approach challenges current assumptions and opens new research paths for alignment.

Plan

We will construct a common mathematical language that bridges the gap between the bottom-up, multilayer architecture and the top-down safety guarantees. Then we will work on theorems and proofs that guarantee the safety and integrity of scale free systems both top-down and bottom-up. Once the theoretical foundations are established, we will develop a framework to evaluate LMAs and run experiments to test our results in representative scenarios, ensuring that the mathematical safety properties hold in practice. The expected outcome is tools and experimental results that validate our formal results, demonstrating that AI systems designed with our framework operate safely even under adverse or unexpected conditions. Our research will involve collaboration with experts in complex systems, biology, and AI. This collaborative effort will ensure that our mathematical abstractions are both theoretically sound and practically relevant. The expected outcome is refined models and proofs that incorporate diverse perspectives and address the challenges of real-world deployment. To reiterate:

• O1 - Formal Foundations: We aim to formally merge the aforementioned lines of research. This will give us a versatile foundation on which to build.

• O2 - Proofs: We will then develop a rigorous, modular framework with formal proofs and methods to to verify safety constraints in distributed, multi-layered, scale-free systems that involve AI.

• O3 - Application: We aim to produce practical tools, and test them against current LMA and simulated scale-free systems that are likely to exist in future.

 

References

[1] Bennett, M.T.: Computational dualism and objective superintelligence. In: 17th International Conference on Artificial General Intelligence. Springer (2024)

[2] Bennett, M.T.: The optimal choice of hypothesis is the weakest, not the shortest. In: 16th International Conference on Artificial General Intelligence, pp. 42–51. Springer (2023)

[3] Friston, K.J., Heins, C., Verbelen, T., Costa, L.D., Salvatori, T., Markovic, D., Tschantz, A., Koudahl, M.T., Buckley, C.L., Parr, T.: From pixels to planning: scale-free active inference. CoRR abs/2407.20292 (2024)

[4] Sol´e, R., Seoane, L.F.: Evolution of brains and computers: The roads not taken. Entropy 24(5), 665 (2022)

[5] Sol´e, R., Kempes, C.P., Corominas-Murtra, B., De Domenico, M., Kolchinsky, A., Lachmann, M., Libby, E., Saavedra, S., Smith, E., Wolpert, D.: Fundamental constraints to the logic of living systems. Interface Focus 14(5), 20240010 (2024)

[6] Man, K., Damasio, A.R.: Homeostasis and soft robotics in the design of feeling machines. Nature Machine Intelligence 1, 446–452 (2019)

[7] Ororbia, A., Friston, K.: Mortal Computation: A Foundation for Biomimetic Intelligence (2024). https://arxiv.org/abs/2311.09589

[8] McMillen, P., Levin, M.: Collective intelligence: A unifying concept for integrating biology across scales and substrates. Communications Biology 7(1), 378 (2024)

[9] Levin, M.: Bioelectrical approaches to cancer as a problem of the scaling of the cellular self. Progress in Biophysics and Molecular Biology 165, 102–113 (2021). Cancer and Evolution

[10] Bennett, M.T.: Are biological systems more ‘intelligent’ than artificial intelligence? (2025)

[11] Leike, J., Hutter, M.: Bad universal priors and notions of optimality. Proceedings of The 28th COLT, PMLR, 1244–1259 (2015)

[12] Davies, P.C.W., Lineweaver, C.H.: Cancer tumors as metazoa 1.0: tapping genes of ancient ancestors. Physical Biology 8(1) (2011)

[13] Elija Perrier, M.T.B.: Position: Stop acting like language model agents are normal agents (2025)

[14] Elija Perrier, M.T.B.: Measuring agency of language model agents (2025)

[15] Maes, P.: Agents that reduce work and information overload. Communications of the ACM 37(7), 30–40 (1994) https://doi.org/10.1145/176789.176792 . Accessed 2024-06-17

[16] Maes, P.: Artificial life meets entertainment: lifelike autonomous agents. Communications of the ACM 38(11), 108–114 (1995)

[17] Lieberman, H.: Autonomous interface agents. In: Proceedings of the ACM SIGCHI Conference on Human Factors in Computing systems. CHI ’97, pp. 67–74. Association for Computing Machinery, New York, NY, USA (1997).

[18] Jennings, N.R., Sycara, K., Wooldridge, M.: A roadmap of agent research and development. Autonomous agents and multi-agent systems 1, 7–38 (1998). Springer

[19] Johnson, D.G.: Software Agents, Anticipatory Ethics, and Accountability. In: Marchant, G.E., Allenby, B.R., Herkert, J.R. (eds.) The Growing Gap Between Emerging Technologies and Legal-Ethical Oversight: The Pacing Problem, pp. 61–76. Springer, Dordrecht (2011). 

[20] Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. The MIT Press, Cambridge, Massachusetts (2018). 

[21] Russell, S.J., Norvig, P.: Artificial Intelligence: A Modern Approach, 4th edn. (2021)

[22] Chan, A., Salganik, R., Markelius, A., Pang, C., Rajkumar, N., Krasheninnikov, D., Langosco, L., He, Z., Duan, Y., Carroll, M., Lin, M., Mayhew, A., Collins, K., Molamohammadi, M., Burden, J., Zhao, W., Rismani, S., Voudouris, K.,

Bhatt, U., Weller, A., Krueger, D., Maharaj, T.: Harms from Increasingly Agentic Algorithmic Systems. In: Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. FAccT ’23, pp. 651–666. Association for Computing Machinery, New York, NY, USA (2023).

[23] Wu, Q., Bansal, G., Zhang, J., Wu, Y., Li, B., Zhu, E., Jiang, L., Zhang, X., Zhang, S., Liu, J., Awadallah, A.H., White, R.W., Burger, D., Wang, C.: AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation Framework. (2023).

[24] OpenAI: OpenAI Charter (2018). https://openai.com/charter/

[25] Gabriel, I. et. al.: The Ethics of Advanced AI Assistants. arXiv. arXiv:2404.16244 [cs] (2024). http://arxiv.org/abs/2404.16244 Accessed 2024-04-26

[26] Kolt, N.: Governing AI Agents, Rochester, NY (2024). 

[27] Lazar, S.: Frontier AI Ethics: Anticipating and Evaluating the Societal Impacts of Generative Agents. arXiv. arXiv:2404.06750 [cs] (2024).

[28] Anthropic: Anthropic’s responsible scaling policy (2023).

[29] Fields, C., Levin, M.: Scale-free biology: Integrating evolutionary and developmental thinking. BioEssays 42 (2020)

Existing resources

We have both made substantial strides in this research independent of this funding, and have complementary projects which will help us realise the vision we describe here.

Open Source Licensing

MIT - Massachusetts Institute of Technology License

Links and references

A full list of 29 references are available in the long description.

Preprints of the two most relevant papers on which this project is based are available here: 

  1. https://osf.io/preprints/osf/e6fky_v2
  2. https://osf.io/3tpq4_v1 

Was there any event, initiative or publication that motivated you to register/submit this proposal?

A personal referral

Proposal Video

Placeholder for Spotlight Day Pitch-presentations. Video's will be added by the DF team when available.

  • Total Milestones

    3

  • Total Budget

    $50,000 USD

  • Last Updated

    23 Feb 2025

Milestone 1 - Experiments

Description

Experiments measuring the agentic identity of language model agents and an accompanying paper.

Deliverables

A research paper detailing our experiments with code.

Budget

$20,000 USD

Success Criterion

The paper should be of a high enough standard to be published at a good venue, and should provide sufficient detail to reproduce the experiments.

Milestone 2 - Integration

Description

Integration of measures of agency with the pancomputational enactivism formalism and formal proofs regarding properties of agency.

Deliverables

A paper unifying these two lines of research detailing theoretical results for agent stability.

Budget

$20,000 USD

Success Criterion

The paper should be of a high enough standard to be published in a good venue.

Milestone 3 - Open Source Tool

Description

An open source tool to measure agent identity that can easily be applied to current language model agents.

Deliverables

A software tool.

Budget

$10,000 USD

Success Criterion

The tool should work and be relatively easy to use.

Join the Discussion (0)

Expert Ratings

Reviews & Ratings

    No Reviews Avaliable

    Check back later by refreshing the page.

feedback_icon