Can we instill heuristics into AI to lock down some dangerous routes of thinking? For example, can we make it assume that “thinking about microbiology or nanotech do not lead to anything interesting” or “if I make a copy of me it will be hostile to me”?
Can we instill heuristics into AI to lock down some dangerous routes of thinking? For example, can we make it assume that “thinking about microbiology or nanotech do not lead to anything interesting” or “if I make a copy of me it will be hostile to me”?