I first want to note that your comment implies my post is saying things which I don’t think it does (or at least I didn’t mean it to): - I didn’t argue that language models are internally human-like, only externally. I do think however that compositional AI systems made of language models should be internally human-like. - I didn’t advocate for training AGI end-to-end (but rather taking the factored cognition approach).
I agree with you that a human-ish-level AGI would be VERY dangerous regardless of how aligned it is (mostly because there are 8 billion people on this planet, and it’s sufficient for just a few of them to use it in stupid/evil ways to put us all in danger).
Thank you for this comment!
I first want to note that your comment implies my post is saying things which I don’t think it does (or at least I didn’t mean it to):
- I didn’t argue that language models are internally human-like, only externally. I do think however that compositional AI systems made of language models should be internally human-like.
- I didn’t advocate for training AGI end-to-end (but rather taking the factored cognition approach).
I agree with you that a human-ish-level AGI would be VERY dangerous regardless of how aligned it is (mostly because there are 8 billion people on this planet, and it’s sufficient for just a few of them to use it in stupid/evil ways to put us all in danger).