I like the general direction of LLMs being more behaviorally “anthropomorphic”, so hopefully will look into the LLM alignment links soon :-)
The useful technique is...
Agree—didn’t find a handle that I understand well enough in order to point at what I didn’t.
We have here a morally dubious decision
I think my problem was with sentences like that—there is a reference to a decision, but I’m not sure whether to a decision mentioned in the article or in one of the comments.
the scenario in this thread
Didn’t disambiguate it for me though I feel like it should.
I am familiar with the technical LW terms separately, so Ill probably understand their relevance once the reference issue is resolved.
there is a reference to a decision, but I’m not sure whether to a decision mentioned in the article or in one of the comments
The decision/scenario from the second paragraph of this comment to wreck civilization in order to take advantage of the chaos better than the potential competitors. (Superhuman hacking ability and capability to hire/organize humans, applied at superhuman speed and with global coordination at scale, might be sufficient for this, no physical or cognitive far future tech necessary.)
didn’t find a handle that I understand well enough in order to point at what I didn’t
The technique I’m referring to is to point at words/sentences picked out intuitively as relatively more perplexing-to-interpret, even without an understanding of what’s going on in general or with those words, or a particular reason to point to those exact words/sentences. This focuses the discussion, doesn’t really matter where. Start with the upper left-hand brick.
I like the general direction of LLMs being more behaviorally “anthropomorphic”, so hopefully will look into the LLM alignment links soon :-)
Agree—didn’t find a handle that I understand well enough in order to point at what I didn’t.
I think my problem was with sentences like that—there is a reference to a decision, but I’m not sure whether to a decision mentioned in the article or in one of the comments.
Didn’t disambiguate it for me though I feel like it should.
I am familiar with the technical LW terms separately, so Ill probably understand their relevance once the reference issue is resolved.
The decision/scenario from the second paragraph of this comment to wreck civilization in order to take advantage of the chaos better than the potential competitors. (Superhuman hacking ability and capability to hire/organize humans, applied at superhuman speed and with global coordination at scale, might be sufficient for this, no physical or cognitive far future tech necessary.)
The technique I’m referring to is to point at words/sentences picked out intuitively as relatively more perplexing-to-interpret, even without an understanding of what’s going on in general or with those words, or a particular reason to point to those exact words/sentences. This focuses the discussion, doesn’t really matter where. Start with the upper left-hand brick.