Well said. I mostly agree, but I’ll note that safety-without-friendliness is good as a non-ultimate goal.
Re human in the loop, I mostly agree. Re situational awareness, I mostly agree and I’ll add that lack-of-situational-awareness is sometimes a good way to deprive a system of capabilities not relevant to the task it’s designed for—“capabilities” isn’t monolithic.
I think my list is largely bad. I think central examples of good-ideas include LM agents and process-based systems. (Maybe because they’re more fundamental / architecture-y? Maybe because they’re more concrete?)
Looking forward to your future-comment-with-suggestions.
Well said. I mostly agree, but I’ll note that safety-without-friendliness is good as a non-ultimate goal.
Re human in the loop, I mostly agree. Re situational awareness, I mostly agree and I’ll add that lack-of-situational-awareness is sometimes a good way to deprive a system of capabilities not relevant to the task it’s designed for—“capabilities” isn’t monolithic.
I think my list is largely bad. I think central examples of good-ideas include LM agents and process-based systems. (Maybe because they’re more fundamental / architecture-y? Maybe because they’re more concrete?)
Looking forward to your future-comment-with-suggestions.