forgive me for not reading that whole post right now, but i suspect that an AI may:
act as we’d want but then sharp left turn once it reaches capabilities that the simulation doesn’t have enough compute for but reality does
need more compute than it has access to in the simulation before it starts modifying the world significantly, so we can only observe it being conservative and waiting to get more compute
act nice because it detects it’s in a simulation, eg by noticing that the world it’s in has nowhere near the computational capacity for agents that would design such an AI
commit to acting nice for a long time and then turn evil way later regardless of whether it’s a computation or not, just in case it’s in a simulation
i believe that “this computation is inside another computation!” is very plausibly an easily guessable idea for an advanced AI. if you address these concerns in your post, let me know and i’ll read it in full (or read the relevant sections) and comment on there.
i believe that “this computation is inside another computation!” is very plausibly an easily guessable idea for an advanced AI.
As stated very early in the article, the AI in simboxes do not have the words for even precursor concepts of ‘computation’ and ‘simulation’, as their knowledge is intentionally constrained and shaped to early or pre civ level equivalents. The foundational assumption is brain-like AGI in historical sims with carefully crafted/controlled ontologies.
forgive me for not reading that whole post right now, but i suspect that an AI may:
act as we’d want but then sharp left turn once it reaches capabilities that the simulation doesn’t have enough compute for but reality does
need more compute than it has access to in the simulation before it starts modifying the world significantly, so we can only observe it being conservative and waiting to get more compute
act nice because it detects it’s in a simulation, eg by noticing that the world it’s in has nowhere near the computational capacity for agents that would design such an AI
commit to acting nice for a long time and then turn evil way later regardless of whether it’s a computation or not, just in case it’s in a simulation
i believe that “this computation is inside another computation!” is very plausibly an easily guessable idea for an advanced AI. if you address these concerns in your post, let me know and i’ll read it in full (or read the relevant sections) and comment on there.
as for my writing style, see here.
I guess I need to write a better concise summary.
As stated very early in the article, the AI in simboxes do not have the words for even precursor concepts of ‘computation’ and ‘simulation’, as their knowledge is intentionally constrained and shaped to early or pre civ level equivalents. The foundational assumption is brain-like AGI in historical sims with carefully crafted/controlled ontologies.