RSS

Robert Kralisch

Karma: 175

Hey, I am Robert Kralisch, an independent conceptual/​theoretical Alignment Researcher. I have a background in Cognitive Science and I am interested in collaborating on an end-to-end strategy for AGI alignment.

I am one of the organizers for the AI Safety Camp 2025, working as a research coordinator by evaluating and supporting research projects that fit under the umbrella of “conceptually sound approaches to AI Alignment”.

The three main branches that I aim to contribute to are conceptual clarity (what should we mean by agency, intelligence, embodiment, etc), the exploration of more inherently interpretable cognitive architectures, and Simulator theory.

One of my concrete goals is to figure out how to design a cognitively powerful agent such that it does not become a Superoptimiser in the limit.

AISC 2023, Progress Re­port for March: Team In­ter­pretable Architectures

Apr 2, 2023, 4:19 PM
14 points
0 comments14 min readLW link

Com­men­tary on “AGI Safety From First Prin­ci­ples by Richard Ngo, Septem­ber 2020”

Robert KralischOct 14, 2021, 3:11 PM
3 points
0 comments19 min readLW link