Do we agree that you can implement more or less winning strategies as a member of the species of homo-sapiens, congruent with the utility-concept of ‘making the world a better place’ , and that there is an absolute ranking criterion on how good said strategies are?
Do we agree that a very common failure mode of homo sapiens is statistical biases in their bayesian cognition, and that these biases have clear causal origin in our evolutionary history?
Do we agree that said biases hamper homo sapiens’ ability to implement winning strategies in the general case?
Do we agree that the writings of Eliezer Yudkowsky and the content of this site as a whole describe ways to partially get around these built in flaws of homo sapiens?
I am fairly confident that a close reading of my comments will find the interprentation of ‘rational’ to be synonymous with ‘winning-strategy-implementation’, and ‘bayesian’ to be synonymous with (in the case that it refers to a person) ‘lesswrong-site-member/sequence-implementor/bayes-conspiracist’ or (in the case that it refers to cognitive architectures) ‘bayesian inference’ and I am tempted to edit them as such.
I am nonplussed at your attempt to lull readers into agreeing with you by asking a lot of rhetorical questions. It’d have been less wrong to post just the last paragraph:
I am fairly confident that a close reading of my comments will find the interprentation [sic] of ‘rational’ to be synonymous with ‘winning-strategy-implementation’, and ‘bayesian’ to be synonymous with (in the case that it refers to a person) ‘lesswrong-site-member/sequence-implementor/bayes-conspiracist’ or (in the case that it refers to cognitive architectures) ‘bayesian inference’ and I am tempted to edit them as such.
The missing link in the argument here is how your examples are, in fact, winning strategies. You claimed some superficial resemblance to things in the sequences, and that you did better than some small sample of humans.
I disapprove of this expanded definition of “bayesian” on the basis that it conflates honest mathematics with handwaving and specious analogies. For example, “it all adds up to normality” is merely a paraphrase of the correspondence principle in QM and does not have any particular legislative force outside that domain.
Great. Now you have really confused me.
Do we agree that you can implement more or less winning strategies as a member of the species of homo-sapiens, congruent with the utility-concept of ‘making the world a better place’ , and that there is an absolute ranking criterion on how good said strategies are?
Do we agree that a very common failure mode of homo sapiens is statistical biases in their bayesian cognition, and that these biases have clear causal origin in our evolutionary history?
Do we agree that said biases hamper homo sapiens’ ability to implement winning strategies in the general case?
Do we agree that the writings of Eliezer Yudkowsky and the content of this site as a whole describe ways to partially get around these built in flaws of homo sapiens?
I am fairly confident that a close reading of my comments will find the interprentation of ‘rational’ to be synonymous with ‘winning-strategy-implementation’, and ‘bayesian’ to be synonymous with (in the case that it refers to a person) ‘lesswrong-site-member/sequence-implementor/bayes-conspiracist’ or (in the case that it refers to cognitive architectures) ‘bayesian inference’ and I am tempted to edit them as such.
I am nonplussed at your attempt to lull readers into agreeing with you by asking a lot of rhetorical questions. It’d have been less wrong to post just the last paragraph:
The missing link in the argument here is how your examples are, in fact, winning strategies. You claimed some superficial resemblance to things in the sequences, and that you did better than some small sample of humans.
I disapprove of this expanded definition of “bayesian” on the basis that it conflates honest mathematics with handwaving and specious analogies. For example, “it all adds up to normality” is merely a paraphrase of the correspondence principle in QM and does not have any particular legislative force outside that domain.
I’ll concede the point, partially because I tire of this discourse.