RSS

Tiling Agents

TagLast edit: Jul 16, 2024, 10:54 AM by Mateusz Bagiński

An agent might have the ability to create similar or slightly better versions of itself. These new agents can in turn create similar /​ better versions of themselves, and so on in a repeating pattern. This is referred to as an agent tiling itself.

This leads to the question: How can the original agent trust that these recursively generated agents maintain goals that are similar to the original agent’s objective?

In a deterministic logical system, assuming that all agents will share the same axioms, “trust” arises from being able to formally prove that the conclusions reached by any subsequently generated agents will be true. The possibility to be able to have this form of trust is influenced by Löb’s theorem. The inability to form this trust is called the Löbian obstacle.

See Also: Löbian obstacle, Löbs theorem, Vingean Agents, Vingean Reflection

References :

Tiling Agents for Self-Mod­ify­ing AI (OPFAI #2)

Eliezer YudkowskyJun 6, 2013, 8:24 PM
88 points
259 comments3 min readLW link

Walk­through of the Tiling Agents for Self-Mod­ify­ing AI paper

So8resDec 13, 2013, 3:23 AM
29 points
18 comments21 min readLW link

Vingean Reflec­tion: Reli­able Rea­son­ing for Self-Im­prov­ing Agents

So8resJan 15, 2015, 10:47 PM
37 points
5 comments9 min readLW link

The Löbian Ob­sta­cle, And Why You Should Care

lukemarksSep 7, 2023, 11:59 PM
18 points
6 comments2 min readLW link

Log­i­cal In­duc­tor Tiling and Why it’s Hard

DiffractorJun 14, 2018, 6:34 AM
4 points
0 comments12 min readLW link

Prob­a­bil­is­tic Tiling (Pre­limi­nary At­tempt)

DiffractorAug 7, 2018, 1:14 AM
14 points
8 comments9 min readLW link

Leav­ing MIRI, Seek­ing Funding

abramdemskiAug 8, 2024, 6:32 PM
264 points
19 comments2 min readLW link

Seek­ing Collaborators

abramdemskiNov 1, 2024, 5:13 PM
57 points
15 comments7 min readLW link

Tiling agents with trans­finite para­met­ric polymorphism

SquarkMay 9, 2014, 5:32 PM
6 points
11 comments2 min readLW link

The al­ign­ment sta­bil­ity problem

Seth HerdMar 26, 2023, 2:10 AM
35 points
15 comments4 min readLW link

Para­con­sis­tent Tiling Agents (Very Early Draft)

IAFF-User-4Apr 2, 2015, 7:27 AM
8 points
5 comments1 min readLW link
(github.com)

Lec­ture Series on Tiling Agents #2

abramdemskiJan 20, 2025, 9:02 PM
16 points
0 comments1 min readLW link

The Pando Prob­lem: Re­think­ing AI Individuality

Jan_KulveitMar 28, 2025, 9:03 PM
127 points
13 comments13 min readLW link

Lec­ture Series on Tiling Agents

abramdemskiJan 14, 2025, 9:34 PM
38 points
14 comments1 min readLW link

Vingean Reflec­tion: Open Problems

abramdemskiJul 3, 2015, 6:44 PM
5 points
3 comments5 min readLW link

Higher Di­men­sion Carte­sian Ob­jects and Align­ing ‘Tiling Si­mu­la­tors’

lukemarksJun 11, 2023, 12:13 AM
22 points
0 comments5 min readLW link

Ra­tional Effec­tive Utopia & Nar­row Way There: Mul­tiver­sal AI Align­ment, Place AI, New Ethico­physics… (Up­dated)

ankFeb 11, 2025, 3:21 AM
13 points
8 comments35 min readLW link
No comments.