Absolutely, while the suggestion in the post is just about controlling LLMs, this is obviously helpful to any A(G)I approach that uses LLMs, such as all various scafolding approaches. Being able to be a lot more sure of your LLM-simulated agent’s emotions/motivations/behavior/education level/etc is basically always helpful.
Yes. Particularly if you’re already performing an internal review for helpfulness and accuracy. I’ll think more about how those labels could be applied, and include this, with credit to you, in my next post on aligning language model agents.
Absolutely, while the suggestion in the post is just about controlling LLMs, this is obviously helpful to any A(G)I approach that uses LLMs, such as all various scafolding approaches. Being able to be a lot more sure of your LLM-simulated agent’s emotions/motivations/behavior/education level/etc is basically always helpful.
Yes. Particularly if you’re already performing an internal review for helpfulness and accuracy. I’ll think more about how those labels could be applied, and include this, with credit to you, in my next post on aligning language model agents.
Much appreciated! Also, I’d love to hear if anyone starts working on stuff along these lines (especially if they were interested in working together).