In any case, this “hot-air balloonist vs. archer” (POP!) comparison seems like some sort of an argument ad hominem -type fallacy, and that’s why I reacted with an ad hominem attack about legos and stuff. First of all, ad hominem is a fallacy, and does nothing to undermine my case. It does undermine the notion that you are being rational.
Secondly, if my person is that interesting, I’d say I resemble the mathematician C. S. Peirce more than Ramakrishna. It seems to me mathematics is not necessarily considered completely acceptable by the notion of rationality you are advocating, as pure mathematics is only concerned about rules regarding what you’d call “maps” but not rules regarding what you’d call “territory”. That’s a weird problem, though.
I didn’t intend it as much of an ad hominem, after all both groups in the comparison are so far quite unprepared for the undertaking they’re trying to do. Just trying to find ways to try to describe the cultural mismatch that seems to be going on here.
I understand that math is starting to have some stuff dealing with how to make good maps from a territory. Only that’s inside the difficult and technical stuff like Jaynes’ Probability Theory or Pearl’s Causality, instead of somebody just making a nice new logical calculus with an operator for doing induction. There’s already some actual philosophically interesting results like an inductive learner needing to have innate biases to be able to learn anything.
That’s a good result. However, the necessity of innate biases undermines the notion of rationality, unless we have a system for differentiating the rational cognitive faculty from the innately biased cognitive faculty. I am proposing that this differentiation faculty be rational, hence “Metarationality”.
In the Cartesian coordinate system I devised object-level entities are projected as vectors. Vectors with a positive Y coordinate are rational. The only defined operation so far is addition: vectors can be added to each other. In this metasystem we are able to combine object-level entities (events, objects, “things”) by adding them to each other as vectors. This system can be used to examine individual object-level entities within the context other entities create by virtue of their existence. Because the coordinate system assigns a moral value to each entity it can express, it can be used for decision making. Obviously, it values morally good decisions over morally bad ones.
Every entity in my system is an ordered pair of the form
). Here x and y are propositional variables whose truth values can be −1 (false) or 1 (true). x denotes whether the entity is tangible and y whether it is placed within a rational epistemology. p is the entity. &p is the conceptual part of the entity (a philosopher would call that an “intension”). p is the sensory part of the entity, ie. what sensory input is considered to be the referent of the entity’s conceptual part. A philosopher would call p an extension. a, b and c are numerical values, which denote the value of the entity itself, of its intension, and of its extension, respectively.
The right side of the following formula (right to the equivalence operator) tells how b and c are used to calculate a. The left side of the formula tells how any entity is converted to vector a. The vector conversion allows both innate cognitive bias and object-level rationality to influence decision making within the same metasystem.
In any case, this “hot-air balloonist vs. archer” (POP!) comparison seems like some sort of an argument ad hominem -type fallacy, and that’s why I reacted with an ad hominem attack about legos and stuff. First of all, ad hominem is a fallacy, and does nothing to undermine my case. It does undermine the notion that you are being rational.
Secondly, if my person is that interesting, I’d say I resemble the mathematician C. S. Peirce more than Ramakrishna. It seems to me mathematics is not necessarily considered completely acceptable by the notion of rationality you are advocating, as pure mathematics is only concerned about rules regarding what you’d call “maps” but not rules regarding what you’d call “territory”. That’s a weird problem, though.
I didn’t intend it as much of an ad hominem, after all both groups in the comparison are so far quite unprepared for the undertaking they’re trying to do. Just trying to find ways to try to describe the cultural mismatch that seems to be going on here.
I understand that math is starting to have some stuff dealing with how to make good maps from a territory. Only that’s inside the difficult and technical stuff like Jaynes’ Probability Theory or Pearl’s Causality, instead of somebody just making a nice new logical calculus with an operator for doing induction. There’s already some actual philosophically interesting results like an inductive learner needing to have innate biases to be able to learn anything.
That’s a good result. However, the necessity of innate biases undermines the notion of rationality, unless we have a system for differentiating the rational cognitive faculty from the innately biased cognitive faculty. I am proposing that this differentiation faculty be rational, hence “Metarationality”.
In the Cartesian coordinate system I devised object-level entities are projected as vectors. Vectors with a positive Y coordinate are rational. The only defined operation so far is addition: vectors can be added to each other. In this metasystem we are able to combine object-level entities (events, objects, “things”) by adding them to each other as vectors. This system can be used to examine individual object-level entities within the context other entities create by virtue of their existence. Because the coordinate system assigns a moral value to each entity it can express, it can be used for decision making. Obviously, it values morally good decisions over morally bad ones.
Every entity in my system is an ordered pair of the form
). Here x and y are propositional variables whose truth values can be −1 (false) or 1 (true). x denotes whether the entity is tangible and y whether it is placed within a rational epistemology. p is the entity. &p is the conceptual part of the entity (a philosopher would call that an “intension”). p is the sensory part of the entity, ie. what sensory input is considered to be the referent of the entity’s conceptual part. A philosopher would call p an extension. a, b and c are numerical values, which denote the value of the entity itself, of its intension, and of its extension, respectively.The right side of the following formula (right to the equivalence operator) tells how b and c are used to calculate a. The left side of the formula tells how any entity is converted to vector a. The vector conversion allows both innate cognitive bias and object-level rationality to influence decision making within the same metasystem.
%20\Leftrightarrow%20{%5E{x}_{y}p}_{\frac{%20\textup{min}(m,n)%20}%20{%20\textup{max}(m,n)%20}(m+n)}%20=(%5E{x}_{y}{\&}p%20_n%20,%20{%5E{x}_{y}*p}%20_m%20)))If someone says that it’s just a hypothesis this model works, I agree! But I’m eager to test it. However, this would require some teamwork.