Language model agents do seem like an enormous business opportunity. They’re a way to fund alignment research if you think that’s a worthwile progress/alignment tradeoff. I also agree with that Christiano post arguing that progressing language model agents is net-neutral or positive, for the reasons he gives and more. But it’s highly debatable.
Language model agents do seem like an enormous business opportunity. They’re a way to fund alignment research if you think that’s a worthwile progress/alignment tradeoff. I also agree with that Christiano post arguing that progressing language model agents is net-neutral or positive, for the reasons he gives and more. But it’s highly debatable.
I wish I could get someone to debate it.
I have a whole schpiel on why and how language model agents are the best shot we’ll get at alignment. These arguments seem obvious, and nobody has given me counterarguments that there’s a better shot. So I’ll keep trying to drum up more interest until someone explains to me why I’m getting it wrong.