What do you disagree about? I agree that understanding addition implies that you understand something important about token accounting. I think there’s something about addition that is maybe best learned by studying token accounting or similar (understanding how minds come to practice addition). I also think much of the essence of [addition as addition itself] is best and most easily understood in a more normal way—practicing counting and computing things in everyday life—and *not* by studying anything specifically about Mesopotamian clay token accounting, because relative to much of the essence of addition, historical accounting systems are baroque with irrelevant detail, and are a precursor or proto form of practicing addition, hence don’t manifest the essence of addition in a refined and clear way.
I like your elevator example. I think it’s an open question whether / how amoebas are manifesting the same principles as (human, say) agency / mind / intelligence, i.e. to what extent amoebas are simpler models of the same thing (agent etc.) vs. models of something else (such as deficient agency, a subset of agency, etc.). I mean, my point isn’t that there’s some “amount” that amoebas “are agents” or whatever, that’s not exactly well-defined or interesting, my point is that the reasons we’re interested in agency make human agency much more interesting than amoeba agency, and this is not primarily a mistake; amoebas pretty much just don’t do fictive learning, logical inference, etc., even though if you try you can read into amoebas a sort of deficient/restricted form of these things.
I don’t know. Possibly something, probably nothing.
the essence of [addition as addition itself]…
The “essence of cognition” isn’t really available for us to study directly (so far as I know), except as a part of more complex processes. Finding many varied examples may help determine what is the “essence” versus what is just extraneous detail.
While intelligent agency in humans is definitely more interesting than in amoebas, knowing exactly why amoebas aren’t intelligent agents would tell you one detail about why humans are, and may thus tell you a trait that a hypothetical AGI would need to have.
knowing exactly why amoebas aren’t intelligent agents would tell you one detail about why humans are
Exactly, yeah; I think in the particular case of amoebas the benefit looks more like this, and it doesn’t so much look like amoebas positively exemplifying much that’s key about the kind of agency we’re interested in re/ alignment. Which is why I disagree with the OP.
What do you disagree about? I agree that understanding addition implies that you understand something important about token accounting. I think there’s something about addition that is maybe best learned by studying token accounting or similar (understanding how minds come to practice addition). I also think much of the essence of [addition as addition itself] is best and most easily understood in a more normal way—practicing counting and computing things in everyday life—and *not* by studying anything specifically about Mesopotamian clay token accounting, because relative to much of the essence of addition, historical accounting systems are baroque with irrelevant detail, and are a precursor or proto form of practicing addition, hence don’t manifest the essence of addition in a refined and clear way.
I like your elevator example. I think it’s an open question whether / how amoebas are manifesting the same principles as (human, say) agency / mind / intelligence, i.e. to what extent amoebas are simpler models of the same thing (agent etc.) vs. models of something else (such as deficient agency, a subset of agency, etc.). I mean, my point isn’t that there’s some “amount” that amoebas “are agents” or whatever, that’s not exactly well-defined or interesting, my point is that the reasons we’re interested in agency make human agency much more interesting than amoeba agency, and this is not primarily a mistake; amoebas pretty much just don’t do fictive learning, logical inference, etc., even though if you try you can read into amoebas a sort of deficient/restricted form of these things.
Good advice for learning in general.
I don’t know. Possibly something, probably nothing.
The “essence of cognition” isn’t really available for us to study directly (so far as I know), except as a part of more complex processes. Finding many varied examples may help determine what is the “essence” versus what is just extraneous detail.
While intelligent agency in humans is definitely more interesting than in amoebas, knowing exactly why amoebas aren’t intelligent agents would tell you one detail about why humans are, and may thus tell you a trait that a hypothetical AGI would need to have.
I’m glad you liked my elevator example!
Exactly, yeah; I think in the particular case of amoebas the benefit looks more like this, and it doesn’t so much look like amoebas positively exemplifying much that’s key about the kind of agency we’re interested in re/ alignment. Which is why I disagree with the OP.