Huh. Yeah, that’s an interesting case which yeah, plausibly doesn’t require any eigening. I think the plausibility comes from it being a case where someone can so fully do it from their personal inside view (the immediate calculation and also their belief in how the underlying mathematical operations ought to work).
I don’t think it scales to anything interesting (def not alignment research), but it is conceptually interesting for how I’ve been thinking about this.
No, it’s the central example for what would work in alignment. You have to think about the actual problem. The difficulty of the problem and illegibility of intermediate results means eigening becomes dominant, but that’s a failure mode.
I agree that eigening isn’t the key concept for alignment or other scientific process. Sure you could describe any consensus that way, but they could be either very good or just awful depending on how much valid analysis went into each step of doing that eigening. In a really good situation, progress toward consensus is only superficially describable as eigening. The real progress is happening by careful thinking and communicating. The eigening isn’t happening by reputation but by quality of work. In a bad field, eigening is doing most of the work.
Referring to them both as eigening seems to obscure the difference between good and bad science/theory creation.
Huh. Yeah, that’s an interesting case which yeah, plausibly doesn’t require any eigening. I think the plausibility comes from it being a case where someone can so fully do it from their personal inside view (the immediate calculation and also their belief in how the underlying mathematical operations ought to work).
I don’t think it scales to anything interesting (def not alignment research), but it is conceptually interesting for how I’ve been thinking about this.
No, it’s the central example for what would work in alignment. You have to think about the actual problem. The difficulty of the problem and illegibility of intermediate results means eigening becomes dominant, but that’s a failure mode.
Interesting to consider it a failure mode. Maybe it is. Or is at least somewhat.
I’ve got another post on eigening in the works, I think that might provide clearer terminology for talking about this, if you’ll have time to read it.
I agree that eigening isn’t the key concept for alignment or other scientific process. Sure you could describe any consensus that way, but they could be either very good or just awful depending on how much valid analysis went into each step of doing that eigening. In a really good situation, progress toward consensus is only superficially describable as eigening. The real progress is happening by careful thinking and communicating. The eigening isn’t happening by reputation but by quality of work. In a bad field, eigening is doing most of the work.
Referring to them both as eigening seems to obscure the difference between good and bad science/theory creation.
But yeah if you mean “I don’t think it scales to successfully staking out territory around a grift” that seems right.