Information and the Continuum Hypothesis

Years ago when I was living in Sweden, I discovered what seems to be an original number I called \Omega, which is defined as the logarithm of \aleph_0, i.e., \Omega = \log(\aleph_0). You can read more about it in this paper [1], where I present a few theorems etc., and introduce a machine that actually has storage equal to \Omega bits. The machine would require infinite time to construct, but so does a true UTM, which has an infinite tape. Similarly, the machine described in [1] has countably many states, and therefore, \Omega is the amount of information that can be stored on any machine that has countably many states. As a consequence, \Omega has intuitive physical meaning, and therefore, we have a reason to assume it exists.

I was thinking about the number again while food shopping, and I realized it fits the continuum hypothesis perfectly. I show in [1] that there is no set with a cardinality of \Omega, and that as a logical consequence, \Omega must be something other than a number. The natural answer is that \Omega is a quantity of information. We can think about zero the same way, as 0 = \log(1), which satisfies our intuition, that zero represents no substance. Further, in this case, zero is the amount of information contained in a system that has exactly one state. This is plainly true, since a system that never changes cannot meaningfully store information, and instead simply exists. Similarly, \Omega is not the cardinality of any set, and therefore not a number, but it is nonetheless a meaningful quantity of information. As a consequence, it’s arguable we cannot perform a number of operations a number of times equal to \Omega (see below for more on this topic).

All of that said, \Omega^{\aleph_0} is defined, because that is simply the product \Omega^{\aleph_0} = \Omega \cdot \Omega \cdot \Omega \ldots, taken a countable number of times. Basic assumptions of arithmetic imply the following inequality:

2^{\aleph_0} \leq \Omega^{\aleph_0} \leq \aleph_0^{\aleph_0} = \aleph_1.

Because 2^{\aleph_0} = \aleph_1, it must be the case that \Omega^{\aleph_0} = \aleph_1.

However, taking the logarithm, we have the following:

\aleph_0 \leq \aleph_0 \log(\Omega) \leq \aleph_0 \Omega.

Corollary 1.9 of [1] proves that \Omega is unique, in that there is no other number of bits X \neq \Omega such that X > n, for all n \in \mathbb{N} and X < \aleph_0. Therefore, \log(\Omega) does not exist. As a result, we learn nothing from taking the logarithm, since \Omega is the smallest infinite quantity. However, we can conceive of a countable collection of machines with \Omega bits of storage, and you can therefore easily show that \aleph_0 \cdot \Omega = \aleph_0.

Most interestingly, it must be the case that,

\aleph_0 \leq \aleph_0^{\Omega} \leq \aleph_1.

Therefore, because of the uniqueness of \Omega, if the continuum hypothesis is true, the above inequality is the unique solution. Note that the continuum hypothesis has been proven independent from the standard axioms. Further, note that this would arguably require assuming we can perform an operation \Omega times, which is not an intuitive assumption. You could argue that we already permit the products and powers of real numbers, which are plainly not cardinals, but we can ascribe physical meaning to such processes through approximation using finite decimal rational numbers, which are the ratios of cardinal numbers. In contrast, \Omega has no approximation.

A few other thoughts, n\Omega = \Omega, for all n \in \mathbb{N}, and you can appreciate this by simply reading [1], even though it’s not proven in [1]. However, I realized tonight, it’s arguable that \Omega^2 is not defined. In contrast, 2\Omega is just the number of bits contained in two machines, and more generally, n\Omega is simply the number of bits contained in n machines. It’s not clear how to think about a collection of machines \Omega in number, precisely because there is no set that contains \Omega elements. Therefore, integer powers of \Omega are arguably undefined. However, as noted above, we can conceive of a countable collection of machines, and you can easily show that \aleph_0 \cdot \Omega = \aleph_0. I’m nearly certain I unpacked all this stuff in Sweden, in my notes, but I didn’t include it in [1], because I was working on completing Black Tree AutoML. That said, though I doubt anyone has done any work on this in the interim anyway, I discovered this stuff around 2018.

Note on Ramsey’s Theorem

It’s always bothered me that Ramsey’s Theorem is not probabilistic. For example, R(3,3), i.e., the smallest order complete graph that contains either a complete graph on 3 vertices, or an empty graph on 3 vertices, is 6. This means that literally every graph with 6 or more vertices contains either a complete graph on e vertices, or an empty graph on e vertices. This is not probabilistic, because it’s simply true, for all graphs on 6 or more vertices. But it just dawned on me, you can construct a probabilistic view of this fact, which is that on fewer than 6 vertices, the probability is less than one, whereas with 6 or more vertices, the probability is 1. This is true in the literal sense, since less than all graphs with fewer than 6 vertices have a complete graph on 3 vertices, or an empty graph on 3 vertices, but some will. I think this could actually be quite deep, and connect to random graphs, but I need some time to think about it.

Another thought, that I think I’ve expressed before, if we can analogize Ramsey’s Theorem to time, then it would imply that certain structures eventually become permanent. This is a truly strange idea, and though I’m just brain storming, intuitively, it doesn’t sound wrong. And now that I’ve thought a bit more about it, I’ve definitely had this idea before:

Specifically, correlation between two random variables can be thought of as an edge between two vertices, where the vertices represent the variables, and the edge represents the presence or absence of correlation. If we consider all random variables together, then it’s clear that having no correlation at all would correspond to an empty graph, and correlation between all variables would correspond to a complete graph. If all graphs are equally likely, no correlation, and total correlation would be equally likely, and in fact the least likely possibilities for any graph with more than two vertices (when compared to at least some but less than total correlation). As a result, if we randomly select, random variables, we should generally find at least some correlation, regardless of their nature or apparent relationships.

If we imagine time quantized on a line, with a vertex representing a moment, and allow for one moment in time to be related to another moment in time by connecting them with an edge, we will have a graph, that just happens to be visualized along a line. Applying Ramsey Theory, we know that certain structures must emerge over time, since we are allowing for the possibility of ever larger graphs. At the same time, the correlation argument above implies that each moment should have some possibly non-causal connection to other moments, producing non-empty graphs. That is, if one moment is connected to another in the remote past, it’s really not credible that it’s causal, and is instead an artifact of this line of thinking. This argument as a whole implies the possibility that reality has non-causal relationships over time, regardless of whether or not the past, present, or future, is memorialized in any way, and regardless of whether or not the past, present, or future is physically real, because these are immutable, abstract, arguments. All of that said, this is a lot to think about, and I need to organize it a bit more, but the core idea seems sound, and that’s disturbing.