Knut Sondre Sæbø
@knut-sondre-saebo·Joined Sep 2024·Ideas
Just pointing out that language tends to describe things as having properties. For example, "the flag is red." But that isn't really accurate; it's more that we perceive the flag as red. The flag doesn't actually possess the property of redness.
Just pointing out that language tends to describe things as having properties. For example, "the flag is red." But that isn't really accurate; it's more that we perceive the flag as red. The flag doesn't actually possess the property of redness.
Just pointing out that language tends to describe things as having properties. For example, "the flag is red." But that isn't really accurate; it's more that we perceive the flag as red. The flag doesn't actually possess the property of redness.
Just pointing out that language tends to describe things as having properties. For example, "the flag is red." But that isn't really accurate; it's more that we perceive the flag as red. The flag doesn't actually possess the property of redness.
Just pointing out that language tends to describe things as having properties. For example, "the flag is red." But that isn't really accurate; it's more that we perceive the flag as red. The flag doesn't actually possess the property of redness.
#4952·Dennis HackethalOP, 4 days ago…the subject-predicate structure…
What could grammar have to do with this? A different language that uses different grammar can still make true statements.
By the way, continuing here may not be in your interest because #4930 breaks the criticism chain. If your goal is to refute the notion that ideas can be true, you’ll probably want to connect your next criticism to #4915 somehow, or one of the ideas above it.
Good idea, but one more question first. When you say a different language with different categories could also make true statements, do you mean truth is just any description that maps onto the states of the world? If so, it seems you can have multiple (indefinitely?) different carvings that all give coherent descriptions of those states.
#4950·Dennis HackethalOP, 5 days agoBut to verify absolute truth you would need to know every possible criticism of an idea.
But we don’t need to verify our ideas. As I wrote in #4891, there’s no criterion of truth to tell that an idea is true. But it can still be true.
That's true. This wasn't meant as an argument against realist truth, and it's probably beside the point I'm making anyway. I was just drawing a distinction: an absolute truth can exist, but without a god's eye view we can never know whether our theories correspond to it.
#4939·Dennis HackethalOP, 7 days agoWhen you say it's 100% true that it's raining, "the facts" you correspond to are already facts within that framework, and not reality.
I think of them as facts of reality. I don’t think about ‘frameworks’. I think the idea of frameworks invites relativism.
We don’t need the molecular level for this. Truth is a very simple concept. No need to complicate it.
"Fact about the world" seems too strong to me. There can be many good explanations of the same reality that carve it up differently. Newton's theories still work pretty well, but Einstein's have a more complete mapping onto reality. I agree "it's raining" has something real grounding it. But "rain" as a category, the subject-predicate structure, water as droplets, just seem to be features of our description. My notion of fact might just be wrong. The idea I have in my head when I think of facts is that the concepts we use are definite ontological categories in reality.
#4938·Dennis HackethalOP, 7 days ago[W]e have no way of verifying that our conceptual carvings track or pick out entities and relations in reality. … [This] definitely rules out absolute truth.
I don’t see how it does. That we have no way to verify our theories (“conceptual carvings”) doesn’t rule out absolute truth. It does sound like we have different notions of ‘absolute truth’ in mind. For mine, see #4894.
Ironically, your idea that theories can be “more true than others” rules out absolute truth in the sense that truth leaves absolutely no room for deviation. Absolute truth is a binary: true or false. Nothing in between.
I completely agree with the definition in 4894. But to verify absolute truth you would need to know every possible criticism of an idea. Without a god’s eye view, you can’t know if your ideas are fallible to a criticism you haven’t detected.
A better framing of what I mean might be «closer to truth». If the theories are consistent with more perspectives (big objects, people, small objects etc.), it is closer to truths. Newton’s theory is in that sense closer to truth than Ptolemy’s geocentric theory.
Have some thoughts, which might be way off. But interested in your response. It seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable without detection. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This gives a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be specified, or a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for the criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
Have some thoughts, which might be way off. But interested in your response. It seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable which is hard to detect. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This might give a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be turned into a function, or just a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for measuring the "hard-to-vary" criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
It seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable without detection. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This gives a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be specified, or a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for the criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
Have some thoughts, which might be way off. But interested in your response. It seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable without detection. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This gives a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be specified, or a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for the criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
I'm not a programmer, so the code below is 100% AI-generated. But here is an attempt. If we normalize a theory into the parts that can be put on a computer, the types it uses, the nodes (specific values) it commits to, and the functions between them, we can score the theory by how many of those parts run.
from dataclasses import dataclass
@dataclass
class Item:
name: str
kind: str # "type", "node", or "function"
programmable: bool # does it compile and run?
@dataclass
class Theory:
name: str
items: list[Item]
12345
def score(self) -> float:if not self.items:return 0.0ok = sum(1 for i in self.items if i.programmable)return ok / len(self.items)
def compare(a: Theory, b: Theory) -> None:
print(f"{a.name:<20} {a.score():.0%}")
print(f"{b.name:<20} {b.score():.0%}")
Demeter theory
demeter = Theory("Demeter", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Goddess", "type", False),
Item("Emotion", "type", False),
Item("Demeter", "node", False),
Item("emotionstate", "node", False),
Item("emotionat", "function", False),
Item("emotion_weather", "function", False),
])
Axial tilt theory
axialtilt = Theory("Axial tilt", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Angle", "type", True),
Item("Insolation", "type", True),
Item("axialtilt", "node", True),
Item("solarconstant", "node", True),
Item("solarangle", "function", True),
Item("insolation_at", "function", True),
Item("temperature", "function", True),
])
compare(demeter, axial_tilt)
Output metrics:
Demeter 25%
Axial tilt 100%
If we normalize a theory into the parts that can be put on a computer, the types it uses, the nodes (specific values) it commits to, and the functions between them, we can score the theory by how many of those parts run.
The program goes through each item in a theory, counts how many are marked True, and divides by the total. That fraction is the score of how hard the theory is to vary.
An item is True if it can be put on a computer, either by reusing an existing type (Float, Int) or by defining a new one that compiles. It is False if no working type system can express it. The user fills in the labels; the program just counts.
Demeter: 2 of 8 items program (Latitude and Temperature). Demeter, her emotions, and the functions linking them to weather can't. So the score is 25%.
Axial tilt: 9 of 9 items program. Standard types, measured constants, and functions from standard physics. So the score is 100%.
If we normalize a theory into the parts that can be put on a computer, the types it uses, the nodes (specific values) it commits to, and the functions between them, we can score the theory by how many of those parts actually run.
from dataclasses import dataclass
@dataclass
class Item:
name: str
kind: str # "type", "node", or "function"
programmable: bool # does it compile and run?
@dataclass
class Theory:
name: str
items: list[Item]
12345
def score(self) -> float:if not self.items:return 0.0ok = sum(1 for i in self.items if i.programmable)return ok / len(self.items)
def compare(a: Theory, b: Theory) -> None:
print(f"{a.name:<20} {a.score():.0%}")
print(f"{b.name:<20} {b.score():.0%}")
Demeter theory
demeter = Theory("Demeter", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Goddess", "type", False),
Item("Emotion", "type", False),
Item("Demeter", "node", False),
Item("emotionstate", "node", False),
Item("emotionat", "function", False),
Item("emotion_weather", "function", False),
])
Axial tilt theory
axialtilt = Theory("Axial tilt", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Angle", "type", True),
Item("Insolation", "type", True),
Item("axialtilt", "node", True),
Item("solarconstant", "node", True),
Item("solarangle", "function", True),
Item("insolation_at", "function", True),
Item("temperature", "function", True),
])
compare(demeter, axial_tilt)
Output metrics:
Demeter 25%
Axial tilt 100%
I'm not a programmer, so the code below is 100% AI-generated. But here is an attempt. If we normalize a theory into the parts that can be put on a computer, the types it uses, the nodes (specific values) it commits to, and the functions between them, we can score the theory by how many of those parts run.
from dataclasses import dataclass
@dataclass
class Item:
name: str
kind: str # "type", "node", or "function"
programmable: bool # does it compile and run?
@dataclass
class Theory:
name: str
items: list[Item]
12345
def score(self) -> float:if not self.items:return 0.0ok = sum(1 for i in self.items if i.programmable)return ok / len(self.items)
def compare(a: Theory, b: Theory) -> None:
print(f"{a.name:<20} {a.score():.0%}")
print(f"{b.name:<20} {b.score():.0%}")
Demeter theory
demeter = Theory("Demeter", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Goddess", "type", False),
Item("Emotion", "type", False),
Item("Demeter", "node", False),
Item("emotionstate", "node", False),
Item("emotionat", "function", False),
Item("emotion_weather", "function", False),
])
Axial tilt theory
axialtilt = Theory("Axial tilt", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Angle", "type", True),
Item("Insolation", "type", True),
Item("axialtilt", "node", True),
Item("solarconstant", "node", True),
Item("solarangle", "function", True),
Item("insolation_at", "function", True),
Item("temperature", "function", True),
])
compare(demeter, axial_tilt)
Output metrics:
Demeter 25%
Axial tilt 100%
#4931·Knut Sondre Sæbø, 7 days agoIt seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable without detection. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This gives a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be specified, or a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for the criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
If we normalize a theory into the parts that can be put on a computer, the types it uses, the nodes (specific values) it commits to, and the functions between them, we can score the theory by how many of those parts actually run.
from dataclasses import dataclass
@dataclass
class Item:
name: str
kind: str # "type", "node", or "function"
programmable: bool # does it compile and run?
@dataclass
class Theory:
name: str
items: list[Item]
12345
def score(self) -> float:if not self.items:return 0.0ok = sum(1 for i in self.items if i.programmable)return ok / len(self.items)
def compare(a: Theory, b: Theory) -> None:
print(f"{a.name:<20} {a.score():.0%}")
print(f"{b.name:<20} {b.score():.0%}")
Demeter theory
demeter = Theory("Demeter", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Goddess", "type", False),
Item("Emotion", "type", False),
Item("Demeter", "node", False),
Item("emotionstate", "node", False),
Item("emotionat", "function", False),
Item("emotion_weather", "function", False),
])
Axial tilt theory
axialtilt = Theory("Axial tilt", [
Item("Latitude", "type", True),
Item("Temperature", "type", True),
Item("Angle", "type", True),
Item("Insolation", "type", True),
Item("axialtilt", "node", True),
Item("solarconstant", "node", True),
Item("solarangle", "function", True),
Item("insolation_at", "function", True),
Item("temperature", "function", True),
])
compare(demeter, axial_tilt)
Output metrics:
Demeter 25%
Axial tilt 100%
#3069·Dennis HackethalOP revised 6 months agoMy critique of David Deutsch’s The Beginning of Infinity as a programmer. In short, his ‘hard to vary’ criterion at the core of his epistemology is fatally underspecified and impossible to apply.
Deutsch says that one should adopt explanations based on how hard they are to change without impacting their ability to explain what they claim to explain. The hardest-to-change explanation is the best and should be adopted. But he doesn’t say how to figure out which is hardest to change.
A decision-making method is a computational task. He says you haven’t understood a computational task if you can’t program it. He can’t program the steps for finding out how ‘hard to vary’ an explanation is, if only because those steps are underspecified. There are too many open questions.
So by his own yardstick, he hasn’t understood his epistemology.
You will find that and many more criticisms here: https://blog.dennishackethal.com/posts/hard-to-vary-or-hardly-usable
It seems to me that "hard-to-vary" is itself the criterion that a theory should be as programmable as possible. As you note, the goal of a theory should be to make it as explicit as possible, and a program is explicitness in its most complete form. Any theory with ambiguous components automatically has a breaking point that is changeable without detection. A programmable theory has strict causal relations all the way from the axioms to the prediction, which makes any change to the components detectable. In other words: a theory is hard to vary to the extent that its components and the couplings between them can be specified as a program. If a theory is vague, you cannot tell when it has been varied.
This gives a concrete operationalization. A breaking point is any place in the formalization where the chain stops being programmable: a primitive with no implementable type, a coupling between components that cannot be specified, or a step that requires implicit theories to fill the explanatory gaps. A mathematical theory with no remaining gaps has zero breaking points and is maximally hard to vary. A theory in natural language is already worse, because words carry ambiguity and vary from mind to mind. This does not rule out better and worse theories in natural language, since we can use more or less ambiguous words and relations. But it does create a hierarchy of hard-to-vary explanations, where the share of the explanation that is programmable, or at least unambiguous, forms the basis for the criterion.
This is probably too crude a formalization. But evaluating the two theories of Demeter's emotions and axial tilt as explanations, you could check how much of each is programmable. Detecting seasons is programmable in both cases through temperature and changes in weather. Demeter's emotions and the causal link from them to the weather, which is the entire explanation, are not programmable. In the axial tilt theory, every component is. So on this measure Demeter scores 25% and axial tilt scores 100%.
#4915·Dennis HackethalOP, 9 days agoIn that case, I'm unclear what "100% true" means.
Perfect correspondence with the facts.
For example, if it’s currently raining, and you say it is, then your statement is 100% true.
Would you agree that this notion of truth amounts to truth relative to our conceptual framework? When you say it's 100% true that it's raining, "the facts" you correspond to are already facts within that framework, and not reality.
At the molecular level there are no discrete raindrops, only a continuous distribution of H2O molecules constantly evaporating and condensing, and some of those very molecules are diffusing through the roof into the house, since no material is 100% impermeable to water vapor.
When we search for truth, I think most of us are trying to understand the causal structure of the universe, not just predict it with our own fitted models. This is just a criticism of this notion of truth, which waters the concept down from what I at least think of as truth. Many incompatible theories can fit the same facts without capturing any causality. If you agree that truth is correspondence with reality, and not with the facts within our conceptual framework, the problem reemerges.
A statement carves the world into concepts standing in relations. For it to correspond with reality, those concepts must pick out genuine entities and relations in reality. But we have no way of verifying that our conceptual carvings track or pick out entities and relations in reality.
You might disagree. But when we search for truth, I think most of us are trying to understand the causal structure of the universe, not just predict it with our own fitted models. This is just a criticism of this notion of truth, which waters the concept down from what I at least think of as truth. Many incompatible theories can fit the same facts without capturing any causality. If you agree that truth is correspondence with reality, and not with the facts within our conceptual framework, the problem reemerges.
A statement carves the world into concepts standing in relations. For it to correspond with reality, those concepts must pick out genuine entities and relations in reality. But we have no way of verifying that our conceptual carvings track or pick out entities and relations in reality. This might not imply that some theories can't be more true than others. But it definitely rules out absolute truth.
When we search for truth, I think most of us are trying to understand the causal structure of the universe, not just predict it with our own fitted models. This is just a criticism of this notion of truth, which waters the concept down from what I at least think of as truth. Many incompatible theories can fit the same facts without capturing any causality. If you agree that truth is correspondence with reality, and not with the facts within our conceptual framework, the problem reemerges.
A statement carves the world into concepts standing in relations. For it to correspond with reality, those concepts must pick out genuine entities and relations in reality. But we have no way of verifying that our conceptual carvings track the world's entities and relations. And therefore we can't know if a statement is true, or if it merely corresponds with the facts (our ideas/perceptions of the world)
When we search for truth, I think most of us are trying to understand the causal structure of the universe, not just predict it with our own fitted models. This is just a criticism of this notion of truth, which waters the concept down from what I at least think of as truth. Many incompatible theories can fit the same facts without capturing any causality. If you agree that truth is correspondence with reality, and not with the facts within our conceptual framework, the problem reemerges.
A statement carves the world into concepts standing in relations. For it to correspond with reality, those concepts must pick out genuine entities and relations in reality. But we have no way of verifying that our conceptual carvings track or pick out entities and relations in reality.
#4906·Dennis HackethalOP, 10 days agoI think you misunderstand both my own argument and the meaning of ambiguity.
You’re saying that, to hold a true idea in the sense of absolute truth in my head, I’d have to have perfect definitions, which require infinite amounts of information, and having all that information is impossible. Right?
While you obviously know what those words mean, you do not have absolute, 100% defined boundaries of what they refer to and what they don't.
I think it’s enough to know what the words mean for the idea to be true. We don’t have to have “100% defined boundaries”.
Truth means correspondence with the facts (Tarski). Not infinite precision.
I think a ‘trick’ cynics use (not maliciously, still I like to call it a trick) is to set an unrealistically high standard for truth. And then, when no idea ends up being able to meet that standard, they say the idea can’t be true.
When we search for truth, I think most of us are trying to understand the causal structure of the universe, not just predict it with our own fitted models. This is just a criticism of this notion of truth, which waters the concept down from what I at least think of as truth. Many incompatible theories can fit the same facts without capturing any causality. If you agree that truth is correspondence with reality, and not with the facts within our conceptual framework, the problem reemerges.
A statement carves the world into concepts standing in relations. For it to correspond with reality, those concepts must pick out genuine entities and relations in reality. But we have no way of verifying that our conceptual carvings track the world's entities and relations. And therefore we can't know if a statement is true, or if it merely corresponds with the facts (our ideas/perceptions of the world)
#4863·Dennis Hackethal, 26 days agoNice work on #4856. Sounds like you’re one of the few who get DD’s stance re creativity.
I don’t think you’re in the Veritula Telegram channel yet. Email me if you want to be: dh@dennishackethal.com
Thanks! Creativity is one of the most interesting ideas in DD's philosophy. If you come across any articles or resources on it that you've found helpful, I'd love for you to send them over.
I'm actually in the channel, just haven't been very active.
I think tractibility lacks the open-ended capacity to reformulate what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
I think this shows that tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
I think tractibility lacks the open-ended capacity to reformulate what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a pre-given frame.
I might be confused about what you mean by tractible. But it seems to me that tractability can't do the work the bounty asks. Tractability is formally defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
I think the core of universal creativity isn't about efficiency, it's the open-ended capacity to restructure what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
I think this shows that tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
I think tractibility lacks the open-ended capacity to reformulate what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
I think this shows that tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
#4856·Knut Sondre Sæbø, 26 days agoI think the core of universal creativity isn't about efficiency, it's the open-ended capacity to restructure what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
I think this shows that tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
An interesting example from cognitive science is the Mutilated Chessboard Problem, which asks whether a board with two same-coloured corners removed can be tiled by dominoes. As a tiling problem the search space is combinatorially explosive. But reframe it as a colour problem and the answer is easy. Every domino covers one black and one white square, and you have unequal numbers of each. The solution came not from searching harder, but from seeing the problem differently.
#4694·Tyler MillsOP revised about 2 months agoBy this standard, a random number generator has universal creativity as well, and is therefore a person. So there must be a standard for personhood other than: able to generate any possible explanation. Such as: can do that tractably.
I think the core of universal creativity isn't about efficiency, it's the open-ended capacity to restructure what counts as a problem, a solution, and relevant data. Creativity is (at least partially) the ability to reformulate the problem space itself, not by ironing out implications of existing theories. An AI and computational systems is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
I think this shows that tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
This also admits of the distinction between AI and AGI (and "universal creativity") as being whether the system is capable of creating knowledge ex nihilo, as argued by Deutsch. Only universal creativity could create knowledge from nothing. Bounded creativity must start with something.
I think DD's view is that creativity is problem-solving at a meta level. True knowledge creation occurs when the problem space itself is reformulated, not by ironing out implications of existing theories. An AI is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
This is why tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
This also admits of the distinction between AI and AGI (and "universal creativity") as being whether the system is capable of creating knowledge ex nihilo, as argued by Deutsch. Only universal creativity could create knowledge from nothing. Bounded creativity must start with something.
Moved the criticism of 4694
#4688·Tyler MillsOP, about 2 months agoThis also admits of the distinction between AI and AGI (and "universal creativity") as being whether the system is capable of creating knowledge ex nihilo, as argued by Deutsch. Only universal creativity could create knowledge from nothing. Bounded creativity must start with something.
This also admits of the distinction between AI and AGI (and "universal creativity") as being whether the system is capable of creating knowledge ex nihilo, as argued by Deutsch. Only universal creativity could create knowledge from nothing. Bounded creativity must start with something.
I think DD's view is that creativity is problem-solving at a meta level. True knowledge creation occurs when the problem space itself is reformulated, not by ironing out implications of existing theories. An AI is already good at ironing out the implications in our language and existing knowledge systems. But that's search within a given space, not the creation of a new one. Creativity seems to work on a higher level. It's operating at the level of problem framing, which requires things like relevance. An AI can't create new relevance, because its weights are a statistical compression of what humans have already found relevant. It inherits a frame; it doesn't generate one.
This is why tractability can't do the work the bounty asks. Tractability is defined relative to a fixed problem space. But universal creativity is (at least partially) the capacity to restructure the space, to change what counts as a problem, a solution, and relevant data.
By Tractible, do you mean "efficient relative to fixed task"?
By Tractible, do you mean "efficient relative to fixed task"?