Before discussing any, let’s quickly recap what ZF (Zermelo-Fraenkel axioms of set theory) and HoTT (Homotopy type theory) are meant to be, shall we?

Part 0: Foundations

So, every mathematical object has it’s own rules; you can talk about groups, probability distributions, matrices, metric spaces and other stuff. It would be nice if all of these things could be described by the same elemental object. A thing to represent everything else.

In this sense, ZF and HoTT are “foundational theories” in the sense that, every structure can be described as an elemental object from the theory.

ZF’s elemental object are the sets. They are meant to capture the idea of a “Collection of things with properties in common”, like the collection of all natural numbers, the collection of all symetries of a triangle, the collection off all the weighted coins, and things like that. Of course, as a couple of you may or may not know, letting everything be a set has a couple of undesirable consequences:

Russell's paradox - Wikipedia
Paradox in set theory In mathematical logic , Russell's paradox (also known as Russell's antinomy ) is a set-theoretic paradox published by the British philosopher and mathematician , Bertrand Russell , in 1901. [ 1 ] [ 2 ] Russell's paradox shows that every set theory that contains an unrestricted comprehension principle leads to contradictions. [ 3 ] According to the unrestricted comprehension principle, for any sufficiently well-defined property , there is the set of all and only the objects that have that property. Let R be the set of all sets that are not members of themselves. (This set is sometimes called "the Russell set".) If R is not a member of itself, then its definition entails that it is a member of itself; yet, if it is a member of itself, then it is not a member of itself, since it is the set of all sets that are not members of themselves. The resulting contradiction is Russell's paradox. In symbols: Let R = { x ∣ x ∉ x } , then R ∈ R ⟺ R ∉ R {\displaystyle {\text{Let }}R=\{x\mid x\not \in x\}{\text{, then }}R\in R\iff R\not \in R} Russell also showed that a version of the paradox could be derived in the axiomatic system constructed by the German philosopher and mathematician Gottlob Frege , hence undermining Frege's attempt to reduce mathematics to logic and calling into question the logicist programme . Two influential ways of avoiding the paradox were both proposed in 1908: Russell's own type theory and the Zermelo set theory . In particular, Zermelo's axioms restricted the unlimited comprehension principle. With the additional contributions of Abraham Fraenkel , Zermelo set theory developed into the now-standard Zermelo–Fraenkel set theory (commonly known as ZFC when including the axiom of choice ). The main difference between Russell's and Zermelo's solution to the paradox is that Zermelo modified the axioms of set theory while maintaining a standard logical language, while Russell modified the logical language itself. The language of ZFC, with the help of Thoralf Skolem , turned out to be that of first-order logic . [ 4 ] The paradox had already been discovered independently in 1899 by the German mathematician Ernst Zermelo . [ 5 ] However, Zermelo did not publish the idea, which remained known only to David Hilbert , Edmund Husserl , and other academics at the University of Göttingen . At the end of the 1890s, Georg Cantor – considered the founder of modern set theory – had already realized that his theory would lead to a contradiction, as he told Hilbert and Richard Dedekind by letter. [ 6 ] Informal presentation [ edit ] Most sets commonly encountered are not members of themselves. Let us call a set "normal" if it is not a member of itself, and "abnormal" if it is a member of itself. Clearly every set must be either normal or abnormal. For example, consider the set of all squares in a plane . This set is not itself a square in the plane, thus it is not a member of itself and is therefore normal. In contrast, the complem

So not every “Collection of any things” can be described as a set. The job of the ZF axioms is to explain which collections are sets and which new sets can be made from previous sets.

HoTT on the other hand is something different. For starters, it’s elemental object are the types. Which encode the idea of “An object with a defined behaviour”. A number behaves in a fundamentally different way than a function, or a group, or a mere set, or a proposition, or a word. You get the idea. While ZF avoids Russell’s paradox by restricting what a set can be, HoTT avoids Russell’s paradox by adding a hierarchy.

You see, the analogous to the “collection of all sets” in type theory isn’t a “member of itself”. The way other theories beyond ZF manage to do this is to declare that the collection of all sets is something higher than a set, is a class. And the collection of all classes would be something even higher than that. This is what I mean by a hierarchy. This would avoid the paradox in a slighltly more convoluted way but it has some advantages like being able to have arbitrarialy large families of sets, classes or things like that.

So the former uses sets (collections of things with shared properties) and the latter uses types (objects with a defined behaviour). What’s the difference?

Part 1: The structure of objects

Let’s consider a group:

Given a set G and an operation [math]\cdot_G : G \times G \to G[/math]. We say that [math](G, \cdot_G)[/math] is a group if the following criteria is met:

  • There is a neutral element on [math]G[/math]. That is, some [math]e \in G[/math] such that [math]e \cdot_G x = x = x \cdot_G e[/math], where [math]x[/math] can be any element on [math]G[/math]
  • The operation is associative, meaning that we don’t have to bother with parenthesis ([math]a \cdot_G b \cdot_G c[/math] isn’t ambiguous because [math](a \cdot_G b) \cdot_G c = a \cdot_G (b \cdot_G c)[/math]).
  • Every object can be inverted. That is, for any [math]x \in G[/math] there is some [math]\bar{x} \in G[/math] such that [math]x \cdot_G \bar{x} = e = \bar{x} \cdot_G x[/math].

An example of a group is the integers with adition. Is easy to verify that all of the criteria are met. Another example are the positive racional numbers with multiplication. Another less “mathy” example are the symmetries of a square (all the things that you can do to a square that leave it looking the same) with the operation of “just make one after the other”.

Notice that a group is just a set (a collection of things) with some added structure (the operation in question. Most mathy things will have the same… feature. A metric space is a collection of things (points) with a distance function. A probability space is a collection of things (the things that could happen) together with a probability distribution (a way to determine the “likelyhood” of the things that could happen). And so on.

HoTT on the other hand, cannot define an object without some behaviour. To make an example, here is how the natural numbers are defined:

Introduction:

  • [math]0 : \N[/math] read as “0 is of type [math]\N[/math]" or “0 has type [math]\N[/math]"
  • [math]succ : \N \to \N[/math], understood as [math]succ[/math] being a function from the natural numbers to the natural numbers. The usual succesor function.

Elimination:

Let [math]P : \N \to U[/math] a family of types indexed by [math]\N[/math]. To show that each [math]P(x)[/math] is inhabited (there is a point [math]a : P(x)[/math] it’s sufficient to show that:

  • [math]P(0)[/math] is inhabited (there is some [math]a : P(0)[/math])
  • Given [math]n : \N[/math], [math]x : P(n)[/math], show a general procedure to construct some inhabitant of [math]P(succ (n))[/math] (Some [math]g[/math] that taking [math]n[/math] and [math]x[/math] produces the inhabitant [math]g(n, x) : P(succ(n))[/math]).

With those two ingredients, there is a function [math]f_{(c, g)}[/math] that takes any [math]x : \N[/math] and produces some [math]f_{(c, g)}(x) : P(x)[/math].

Computation:

The function [math]f_{(c, g)}[/math] behaves as espected:

  • [math]f_{(c, g)}(0) = c : P(0)[/math]
  • [math]f_{(c, g)}(succ(n)) = g(n, f_{(c, g)}(n))[/math]

Reading in between all the jargon, you might notice that the “elimination-computation” parts are just the induction principle for the natural numbers. More over, the computation part looks very similar to the recursive definitions for functions. This is because on HoTT, propositions are also representable by types, so the recursive theorem and the induction principle are two faces of the same coin (the former is a special case of the latter).

If you think about it, everything from [math]\N[/math] can be constructed using it’s recursion and induction. So this type fully captures what [math]\N[/math] is meant to represent. We can also have a universe of groups by considering all the types on a universe (one of the levels from the hierarchy discussed before) which some operation that satisfies all of the criteria required. “An object with a defined behaviour” can encapsulate all sorts of structure, call it distance function, call it order relationship, call it whatever. HoTT allows you to construct a single type which encodes all of this information.

But other than that small detail, there doesn’t seem to be an improvement, does it?

Part 2: Types aren’t sets

A set is just a collection of points (things). There is no aditional information on the collection itself. That’s why it is neccesary to add some aditional structure. The sets have this feature called extensionality where two sets are the same when they have the exact same objects. This is good because it means that the collection of all the natural numbers that can be expressed as a product of two naturals bigger than 1 and the collection of all the natural numbers which aren’t prime nor 1 are the same collections. Regardless of the properties defining them being different.

But it also means that a set “forgets” where it came from. Given an arbitrary collection of things, it can be very hard to find the property (or one of the properties) that they had in common.

I casually threw the following sentence before:

Propositions are also representable by types

This is a huge game changer, because that means that types are intentional, they can “remember” which properties defined them. A member from “the type of all natural numbers which can be expressed as a product of two naturals bigger than 1” is not just a number (like [math]6[/math]) but also a proof of that number being expressable as a product of two naturals bigger than 1 (meaning, a proof of [math]6 = 3 \times 2[/math] and two proofs of [math]3 > 1[/math] and [math]2 > 1[/math]), while a member from “the type of all natural numbers which aren’t prime nor 1” is a number (like [math]6[/math]) together with a proof that it isn’t a prime nor is it [math]1[/math] (meaning, a proof that [math]6[/math] isn’t a prime number, and a proof that [math]6 \ne 1[/math]).

You may think that this means that the former and the latter aren’t equal since their points have very different structures. But the way equality is even treated in types is more… complicated, than it is for sets.

Part 3: Defined or proved?

Taking [math]P: \N \to U[/math] to be constant at [math]\N[/math] ([math]P(x)[/math] is [math]\N[/math] for all [math]x : \N[/math]) we can define a function from [math]\N[/math] to [math]\N[/math] by:

  • [math]f(0) = 1[/math]
  • [math]f(succ (n)) = (succ (n)) \times f(n)[/math]

it doesn’t make any sense at all if someone asked us to “proof that [math]f(0)[/math] is equal to [math]1[/math]". At that question we would answer simply “that’s true by definition”. It may be less obvious if someone asked us to proof that “[math]f(4) = 24[/math]". Simply stating “by definition” and calling it done doesn’t seem as convincing as the previous time, and indeed it isn’t, since in this case the proof requires multiple steps, not just one:

  • [math]f(4) = 4 \times f(3) = 4 \times 3 \times f(2) = \cdots = 4 \times 3 \times 2 \times 1 \times f(0) = 4 \times 3 \times 2 \times 1 \times 1 = 24[/math]

We usually don’t notice it, but equality is a proposition. That is, “[math]a = b[/math]" is just a statement that may or may not hold between two pre-existing objects [math]a[/math] and [math]b[/math]. Or at least, that’s what ZF agrees on. But there is obviously something fundamentally different between defining something and proving, when we state that [math]f(0) = 1[/math] we’re not trying to say “we have a proof that this rare function [math]f[/math] evaluated at [math]0[/math] returns the value [math]1[/math]". Instead, we mean something more along the lines of “we declare that this function at [math]0[/math] takes the value [math]1[/math]”; we declare that [math]f(0)[/math] and [math]1[/math] are exactly the same thing.

It is different to say that something is defined to be something else, at proving that something is something else. For example, we can define [math]\sqrt{2}[/math] to be some positive real number which the property that [math]\sqrt{2} \times \sqrt{2} = \sqrt{2}^2 = 2[/math]. But that definition doesn’t acert that such real number actually exists. We can proof that the number actually exists by constructing the following real number

[math]sup \{x : \R, x^2 < 2\}[/math]

and showing that this real number has indeed the desired property.

There is a fundamental distinction between declaring that something is literally the same thing as something else, and proving that something is indeed like something else. HoTT makes this distinction even more explicit, writting “[math]a \equiv b[/math]" to mean “[math]a[/math] is literally the same thing as [math]b[/math]" and “[math]a = b[/math]" to mean “the proposition that [math]a[/math] is indeed like [math]b[/math]".

With this distinction in mind we can finally understand the types from the previous section. They are not literally the same thing, but you can proof that they are alike.

Final: base = base, and beyond.

by the way that I’ve presented it, it looks like HoTT deals with a lot more of complications than ZF to acomplish the same things. But trust me when I say that there are types that quite literally don’t have an analogous as a set.

Take the following type:

Introduction:

[math]base : S^1[/math]

[math]loop : base =_{S^1} base[/math]

This may look like something absurd, since its a type with a single point (base) and a proof of that point being indeed like itself (loop). But due to the nature of equality ([math]a = b[/math] and [math]b = c[/math] implies [math]a = c[/math], [math]a = b[/math] implies [math]b = a[/math]) this single “non-trivial proof” that base is indeed like base gives rice to a much richer structure.

Taking [math]a[/math], [math]b[/math], [math]c[/math] to be [math]base[/math], we have that [math]loop[/math] is a proof that [math]base = base[/math] and [math]loop[/math] is a proof that [math]base = base[/math]. Then we can construct a new proof [math]loop^2[/math] that ensures the obviousness of [math]base = base[/math]. Likewise, since [math]base = base[/math] ([math]a = b[/math]) we should be able to “reverse” this proof to obtain a new proof [math]loop^{-1}[/math] of the fact that [math]base = base[/math] ([math]b = a[/math]). And applying transitivity with these two proofs should give rice to the trivial proof that [math]base = base[/math].

Indeed, this trivial proof is called reflexivity on [math]a[/math] or [math]refl_a[/math] for short. Is the trivial proof that any object is indeed like itself. And while [math]loop \circ loop^{-1}[/math] may not be exactly the same thing as [math]refl_{base}[/math], they are indeed like eachother.

It turns out that this misterious type is actually a circle. Thinking on base as some arbitrary point, and loop as an actual loop starting and ending at base. You may see why this type reflects (in some way) the structure of a circle. To fully capture the analogous of this using sets it is not enough to have the collection of all points of a circle (since that doesn’t capture where [math]loop^2[/math] or [math]loop^{-1}[/math] come from) but to have some aditional structure. This gives rise to homotopy theory which, broadly speaking, is the study of paths between points, and paths between paths or “homotopies”.

I don’t know how to end this text…

View question
About · Careers · Privacy · Terms · Contact · Languages · Your Ad Choices · Press ·
© Quora, Inc. 2025