Hard to Vary or Hardly Usable?
Showing only #3703 and its comments.
See full discussionLog in or sign up to participate in this discussion.
With an account, you can revise, criticize, and comment on ideas.Deutsch’s stance in my own words:
The distinguishing characteristic between rationality and irrationality is that rationality is the search for good explanations. We make progress by searching for good explanations.
A good explanation is hard to vary “while still accounting for what it purports to account for.” (BoI chapter 1 glossary.) A bad explanation is easy to vary.
For example, the Persephone myth as an explanation of the seasons is easy to change without impacting its ability to explain the seasons. You could arbitrarily replace Persephone and other characters and the explanation would still ‘work’. The axis-tilt explanation of the earth, on the other hand, is hard to change without breaking it. You can’t just replace the axis with something else, say.
The quality of a theory is a matter of degrees. The harder it is to change a theory, the better that theory is. When deciding which explanation to adopt, we should “choose between [explanations] according to how good they are…: how hard to vary.” (BoI chatper 9; see similar remark in chapter 8.)
Deutsch leaves open how we find out how hard to vary an explanation is. We need more details. In some cases it’s obvious, but we need a general description for less-obvious cases.
Even if we allow creative user input, eg a score for the quality of an explanation, we run into all kinds of open questions, such as what upper and lower limits to use for the score, and unexpected behavior, such as criticisms pushing an explanations score beyond those limits.
Deutsch contradicts his yardstick for understanding a computational task. He says that you haven’t understood a computational task if you can’t program it. His method of decision-making based on finding good explanations is a computational task. He can’t program it, so he hasn’t understood it.
Isn’t this basically asking for a specification of the creative program? Isn’t this effectively an AGI project?
Isn’t this asking for a formalization of creativity, which is impossible?
No, it’s asking for a formalization of rational decision-making, which is a related but separate issue. Given a set of explanations (after they’ve already been created), what non-creative sorting algorithm do we use to find the best one?
Deutsch says to choose between explanations “according to how good they are” – note the plural.
What if I can only come up with one explanation? Can I just go with that one? What if it’s bad but still the best I could do? He leaves such questions open.
Deutsch says rationality means seeking good explanations, so without a step-by-step guide on how to seek good explanations, we cannot know when we are being irrational. That’s bad for error correction.
From my article:
[I]sn’t the difficulty of changing an explanation at least partly a property not of the explanation itself but of whoever is trying to change it? If I’m having difficulty changing it, maybe that’s because I lack imagination. Or maybe I’m just new to that field and an expert could easily change it. In which case the difficulty of changing an explanation is, again, not an objective property of that explanation but a subjective property of its critics. How could subjective properties be epistemologically fundamental?