Seems to be partially answered in the article.
Search for "Is There a Better Notation?"
The way I learned boolean algebra was by associating certain operations (AND, NOT, OR, etc) to truth tables. In this framework, proving a theorem of boolean algebra would just involve enumerating all possible truth assignments to each variable and computing that the equation holds.
There is another framework for boolean algebra that does not involve truth tables. This is the axiomatic approach [1]. It puts forth a set of axioms (eg "a OR b = b OR a"). The symbol "OR" is not imbued with any special meaning except that it satisfies the specified axioms. These axioms, taken as a whole, implicitly define each operator. It then becomes possible to prove what the truth tables of each operator must be.
One can ask how many axioms are needed to pin down the truth table for NAND. As you know, this is enough to characterize boolean algebra, since we can define all other operators in terms of NAND. It turns out only one axiom is needed. It is unclear to me whether this was first discovered by Wolfram, or the team of William McCune, Branden Fitelson, and Larry Wos. [2]
[1] https://en.wikipedia.org/wiki/Boolean_algebra_(structure)
[2] https://en.wikipedia.org/wiki/Minimal_axioms_for_Boolean_alg...,.
The former one is set theoretic i.e. set of objects (eg. terms) and operations (eg. and/or/not) defined on those objects. The latter is an algebraic specification where a number of properties (expressed by logical formulas which can be axioms or theorems) are expected to be satisfied by the operations.
Also see https://en.wikipedia.org/wiki/Minimal_axioms_for_Boolean_alg...
PS: I would take any claim by Stephen Wolfram as having been "the first" to discover anything with a boatload of salt.
These days, an LLM will, perhaps.
An make it palatable to puny humans.
We currently pull out our phones at the pub/table to check something someone makes up to see if it's legitimate. Now we've invented the technology to have something be that thing that creates a half-truth from what it has absorbed.
Yeah, I wish he had started by defining that. The is hard to understand without it.
Search for "Is There a Better Notation?" in the article, it seems "." is NAND
NAND isn't THE operator, it is AN operator that can be in that place.
If there's only 1 value that variables can be - "I", then "I \dot I = I" would be a valid operator that follows given axioms
Quote: "the · dot here can be thought of as representing the Nand operation"
Search for "What Actually Is the “·”?" for the answer, it's quite complex and fascinating.
you can use whatever symbol you want
all you really know - is the axiom you're given, which transforms one tree structure of operator application into a different one
NAND suggested in replies isn't THE operator, it's AN operator that follows given rules
Nonetheless, every time I browse through Wolfram Alpha I am thoroughly impressed of what it can do with only Natural Language as an input.
The question "is Stephen Wolfram being taken seriously by the mathematics community?" seems relevant as a question to gauge whether one should spend time reading a very long article.
edit: an even more relevant analogy - is Mochizuki's big ego irrelevant in a discussion about whether his proof of the abc conjecture, that nobody understands and which he refuses to explain properly, is correct?
I've never seen a single of these things being called out.
This is an article about a proof Stephen Wolfram claims to have discovered. Somebody further upthread already mentioned that this may not actually be true, or that the credit at least should be shared.
I've never seen a single of these things being called out.
Happens all the time plus many repetitive things constantly get moderated away by users and moderators. The goal is a less boring messageboard, but it's aspirational rather than perfectly achievable.
Edit: Ok, fine, let's do analogies. Imagine Elon Musk started writing long, interesting, researched articles about some topic, say, Chinese history. If every such article got flooded with discussions about who does and does not like Elon Musk and by how much, those would start getting modbleatings too. This is what happened except it wasn't Elon Musk but Stephen Wolfram and it wasn't Chinese History and it started about a decade ago (as you can see from the comment dates).
I'm not claiming that Stephen Wolfram is an idiot or that nothing he writes is of value, but his ego clearly does influence his judgement so it makes no sense to me to claim that it is off-topic (in the way that, say, his love life or his opinion on abortion would be).
Also this:
> Happens all the time plus many repetitive things constantly get moderated away by users and moderators. The goal is a less boring messageboard, but it's aspirational rather than perfectly achievable.
No sorry, I've been on HN during the pandemic and I was this close to quitting it forever. These things do not get called out in any consistent fashion for many topics. I'm pretty sure a message board for people interested in technology can handle a couple of "Wolfram sucks" comments a year better than literally hundreds of angry rants about a health advisor that doesn't even have any relevance to anyone who isn't from the US.
It's not about the reason, it's about the repetitiveness. The boringness of the repetitiveness is one of the organizing principles of HN. Also, people don't actually 'keep calling out' this stuff - these moderator interventions, specifically in the Wolfram case, have been surprisingly effective. Many Wolfram-y articles have neither moderation intervention or bad tangents.
These things do not get called out in any consistent fashion for many topics.
You don't see everything, neither do other forum participants, including the moderators. The fact that, collectively, some boring and repetitive stuff is missed doesn't imply we can all use more boring, repetitive stuff.
The stuff that I mentioned was really hard to miss.
The question OP asked was, essentially, "is Stephen Wolfram actively contributing to the mathematical community?". If you really think that this is off-topic then... I don't know what to say.
A few years ago Sean Carroll was hosting him on his podcast. It was a bit surprising to me because Sean would never give a crackpot the time of the day, and Wolfram is borderline in crackpot territory IMHO, but not quite. He hasn't published anything meaningful in a scientific journal in a long time as far as I know.
* Ref: https://www.wolframphysics.org/technical-introduction/ The podcast with Sean Carroll that parent mentioned is also a surprisingly accessible lay introduction, definitely worth a listen.
Like all the stuff about Rule 30 that he writes _endlessly_ about is both true and about 1/100th as interesting as he evidently thinks it is.
A lot of mathematics is about exploration. You look at some system and try to figure out how it works. If you have a good idea (conjecture), you try proving or disproving it. The goal is to gain some understanding.
Once in a while, it turns out that exploration hits the gold ore in the mine. You get something that applies to solving some problem we've had for years. As an example, algebra was considered meaningless for years. Then cryptography came along.
There are other good examples. Reed-Solomon coding relies on a basis of finite fields.
The problem is we don't really know when we'll strike gold in many cases. So people go exploring. It also involves humans, and they can disagree on a lot of things. Stephen seems to run into disagreements more often than the average researcher, at least historically.
It's like rewriting (and clarifying things here or there, putting in new analogies, etc.. not just translating) everything in Klingon, and creating new Klingon words while he is doing that.
Sure this is "interesting" in academic sense. Good luck finding a journal accept paper written in Klingon.
I think, specifically, his vision is very different and radical from other things I've seen, which you'd think would lend itself to at least one currently-testable hypothesis, but I haven't seen one. It seems there may overfitting. He has a way for every major principle in physics to fit into his theory, but no new ones, or flaws in current ones; that's what worries me. I hope he continues working on it. (Unlikely he'll stop after this many decades)
As a comparison, I'm more hopeful for this than String theory, and the latter has been found worthy of physicist's time, so I certainly think this is.
Regarding the crackpot angle: I think that is not an appropriate label: If you look at the people that the label canonically applies to, you will find people who have no idea what they're talking about, who don't understanding the math and physics etc. That is not the case here; he knows it as well as anyone.
As far as I can tell he's not upset by lack of recognition from the scientific community and he puts his ideas out there for criticism so it seems like he's making a valuable contribution. I don't think scientists should be pre-judged on the perceived importance of their work as that leads to group-think and a lack of exploration.
Metamath: https://us.metamath.org/mpeuni/mmset.html#axioms
Isabelle/ZF: https://isabelle.in.tum.de/dist/library/FOL/ZF/ZF_Base.html
So the "C" from ZFC is "challenging" for type theory? Because, and I think I am not that much overshooting, fundamental axioms are really, really simple, even the "C" from ZFC, and if is "challenging" to express it using some fundamental theory, well, not good omens for that theory, or they can rebuild the realm of classic maths with their own "C"??
most mathematicians will eventually agree with that when pressed. However, almost none of them know the axioms of ZFC by heart (because they don't need them). You can swap out ZFC for something else and nobody will care very much as long as "the same" (does a lot of work here) theorems, which they know and use it their work, remain true.
This is what many theorem provers do. Many aren't based on set theory, for example, but instead on type theory. (You can still do set theory in a framework based on type theory, and vice versa, but the foundations are "different").
Coq, Agda, and Lean use Type Theory.
Generally, most computer scientists using proof automation don't care about the proof itself -- they care that one exists. It can contain as many lemmas and steps as needed. They're unlikely to ever read it.
It seems to me that LLMs would be decent at generating proofs this way: so long as they submit their tactics to the proof checker and the proof is found they can generate whatever is needed.
However for mathematicians, of which I am not a member of that distinguished group, seem to appreciate qualities in proofs such as elegance and simplicity. Many mathematicians that I've heard respond to the initial question believe that a proof generated by some future AI system will not be useful to humans if they cannot understand and appreciate it. The existence of a proof is not enough.
Now that we're getting close to having algorithms that can generate proofs it makes the question a bit more urgent, I think. What use is a proof that isn't elegant? Are proofs written for a particular audience or are they written for the result?
A proven conjecture is IMO better than an unproven one.
But usually, a new proof would shed new lights or build bridges between concepts that were before unrelated.
And in that sense, a proof not understandable by humans is disappointing, because it doesn't really fullfil the need to understand the reason behind why it's true.
Temporarily, Germans won.
Personally, I don't think that proofs that cannot be understood have no value. We rely on such proofs all the time in our day-to-day interpretation of the world around us, our ability to navigate it and anticipate it. I.e. there's some sort of an automated proof tool in our brains that takes the visual input, feeling of muscle tonus, feeling of the force exerted on our body etc. and then gives an answer as to whether we are able to take the next step, pick up a rock and so on.
But, mathematicians also want proofs to be useful to explain the nature of the thing in question. Because another thing we want to do about things like picking up rocks, is we want to make that more efficient, make inanimate systems that can pick up rocks etc.
----
NB. I'm not entirely sure how much LLMs can contribute in this field. The first successes of AI were precisely in the field of automated proofs, and that's where symbolic AI seems to work great. But, I'm not at all an expert on LLMs. Maybe there's some way I cannot think about that they would be better at this task, but on the face of it they just aren't.
I suspect it is the latter that will suffer in automated proofs perhaps - without understanding the techniques, or if the technique is not really novel but just tedious.
Part of it is, I think, that "elegance" is flowery language that hides what mathematicians really want: not so much new proofs as new proof techniques and frameworks. An "elegant" proof can, with some modification, prove a lot more than its literal statement. That way, even if you don't care much about the specific result, you may still be interested because it can be altered to solve a problem you _were_ interested in.
1: It doesn't have to be as big of a deal as this.
Agree. The truthfulness of the four-colour theorem is good to know, although there is not yet any human-readable proof.
For a general theorem prover, reading the code doesn't explain anything insightful about why the theorem is true. For the 4 color theorem, the code that proved it actually gives insight in how the proof works.
The article explains that two mathematicians were able to place Apery's proof that ζ(3) is irrational into a much wider (and hence more powerful) framework. I doubt that framework is as easy to understand as the original proof. But in the end something with wider applicability did come out of the proof.
Still, the newer methods publicized in the Quanta article definitely get more involved, and at least from my perspective they don't establish things as elegantly as Apéry's ζ(2) and ζ(3) arguments do. Hopefully they turn out to be powerful in practice, to make up for it.
That survives in the culture of mathematics where we continue to see a high regard for truth, beauty, and goodness. Which, incidentally, are directly related to logic, aesthetics, and ethics.
The value of truth in a proof is most obvious.
The value of aesthetics is harder to explain, but there's no denying that it is in fact observably valued by mathematicians.
As for ethics, remember that human morality is a proper subset thereof. Ethics concerns itself with what is good. It may feel like a stretch, but it's perfectly reasonable to say that for two equally true proofs of the same thing, the one that is more beautiful is also more good. Also, obviously, given two equally beautiful proofs, if only one is true then it is also more good.
As a non-mathematician, I've noticed this as well, and I have a suspicion the historical "culture" is holding the field back. Gödel proved there are an infinite number of true arithmetic statements unprovable within any (consistent, sufficiently powerful) formal system. But our "gold standard" formal system, ZFC, has about as many axioms as we have fingers — why is finding more axioms not the absolute highest priority of the field?
We struggle to prove facts about Turing machines with only six states, and it's not obvious to me that ZFC is even capable of resolving all questions about the behavior of six state Turing machines (well, specifically just ZF, as C has no bearing on these questions).
Yet Turing machines are about as far from abstract mathematics as one can get, because you can actually build these things in our physical universe and observe their behavior over time (except for the whole "infinite tape" part). If we can't predict the behavior of the majority of tiny, deterministic systems with ZFC, what does that say about our ability to understand and predict real world data, particularly considering that this data likely has an underlying algorithmic structure vastly more complex than that of a six state Turing machine?
More formally, my complaint with the culture of mathematics is:
1) We know that for any string of data, I(data : ZFC) ≤ min(K(data), K(ZFC)) + O(1)
2) K(ZFC) is likely no more than a few bytes. I think the best current upper bound is the description length of a Turing machine with a few hundred states, but I suspect the true value of K(ZFC) is far lower than that
3) Thus K(data) - K(data | ZFC) ≤ "a few bytes"
Consider the massive amounts of data that we collect to train LLMs. The totality of modern mathematics can provide no more than a few bytes of insight into the "essence" of this data (i.e., the maximally compressed version of the data). Which directly translates to limited predictability of the data via Solomonoff induction. And that's in principle — this doesn't even consider the amount of time involved. If we want to do better, we need more axioms, full stop.
One might counter, "well sure, but mathematicians don't necessarily care about real world problems". Ok, just apply the same argument to the set of all arithmetic truths. Or the set of unprovable statements in the language of a formal system (that are true within some model). That's some interesting data. Surely ZFC can discover most "deep" mathematical truths? Not very likely. The deeper truths tend to occur at higher levels of the arithmetic hierarchy. The higher in the hierarchy, the more interesting the statement. And these are tiny statements too: ∀x ∃y ∀z [...]. Well we're already in trouble because ZFC can only decide a small fraction of the Π_2 statements that can fit on a napkin and it drops off very quickly at higher levels than that. Again, we need more axioms.
The infinite tape part isn't some minor detail, it's the source of all the difficulty. A "finite-tape Turing machine" is just a DFA.
Oh is that all? If resource bounded Kolmogorov complexity is that simple, we should have solved P vs NP by now!
I debated adding a bunch of disclaimers to that parenthetical about when the infinite tape starts to matter, but thought, nah, surely that won’t be the contention of the larger discussion point here haha
> As a non-mathematician, I've noticed this as well, and I have a suspicion the historical "culture" is holding the field back.
Holding the field back from what? If the goal of the practitioners of the field is to seek mathematically beauty, then well, that is what they will focus on.
Besides that, I don't really follow your argument about Godel & information theory & that adding more axioms is the key to moving math forwards. In the vast majority of cases, the difficulty in finding a proof of a statement is not that the statement isn't provable under a given formal system, it's that we simply can't find it. But maybe I misunderstand you?
The idea that there might be more axioms that would let one deduce more about computable complexity classes or the like seems pretty unlikely.
The number of provable statements and unprovable statements is countably infinite and we aren’t lacking the ability to prove things due to obviously true missing axioms.
I suggest reading Kanamori's book [3] if you want to learn more about this direction. (There are plenty of recent developments in the field which occured after the publication of that book - for an example of cutting edge research into new axioms, see the paper [4] mentioned in one of the answers to [5].)
If you are only interested in arithmetic consequences of the new axioms, and if you feel that consistency statements are not too interesting (even though they can be directly interpreted as statements about whether or not certain Turing machines halt), you should check out some of the research into Laver tables [6], [7], [8], [9]. Harvey Friedman has also put a lot of work into finding concrete connections between advanced set-theoretic axioms and more concrete arithmetic statements, for instance see [10].
[1] https://mathoverflow.net/questions/479079/why-believe-in-the... [2] https://cs.nyu.edu/pipermail/fom/2000-January/003643.html [3] "The Higher Infinite: Large Cardinals in Set Theory from their Beginnings" by Akihiro Kanamori [4] "Large cardinals, structural reflection, and the HOD Conjecture" by Juan P. Aguilera, Joan Bagaria, Philipp Lücke, https://arxiv.org/abs/2411.11568 [5] https://mathoverflow.net/questions/449825/what-is-the-eviden... [6] https://en.wikipedia.org/wiki/Laver_table [7] "On the algebra of elementary embeddings of a rank into itself" by Richard Laver, https://doi.org/10.1006%2Faima.1995.1014 [8] "Critical points in an algebra of elementary embeddings" by Randall Dougherty, https://arxiv.org/abs/math/9205202 [9] "Braids and Self-Distributivity" by Patrick Dehornoy [10] "Issues in the Foundations of Mathematics" by Harvey M. Friedman, https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3137697
What are I and K? What does "data : ZFC" mean?
Do you feel this could be a matter of framing? If you view the "proof" as being the theorem prover itself, plus the proof that it is correct, plus the assumptions, then whatever capability it gains that lets it prove your desired result probably is generalizable to other things you were interested in. It would seem like a loss if they're dismissed simply because their scratch work is inscrutible.
What I mean is that one could choose a particular tactic to proving a lemma or even rely on certain lemma where that choice is intended for the audience the author has in mind in order to help them understand the steps better.
The context an LLM uses is limited (though growing). However its context is lacking the ability to understand the mind of the interlocutor and the proof they would expect or find helpful.
"Why this proof?" also has a more global context as well. It seems there are have been styles and shifts in cultural norms over the years in proofs. Even though the Elements stood up until the 19th century we don't necessarily write proofs in the style of Euclid even though we may refer to those themes and styles, on purpose, in order to communicate a subtle point.
We have already accepted some proofs we cannot fully understand, such as the proof of the four color theorem that used computational methods to explore a large solution space and demonstrate that no possible special-case combinations violate the theorem. But that was just one part of the proof.
I wonder what we know about proof space generally, and if we had an ASI that reasoned in a substantially different way than humans, what types of proofs it would be likely to generate. Do most proofs contain structural components that humans find pleasing? Do most devolve into convoluted case analyses? Is there a simplest form that a set of correct proofs could be reduced to?
There will perhaps need to be a transition period where we might need to look at basic type theory augmenting or replacing material in introductory proof classes. Instead of truth tables and ZFC, teach math-as-programming. Propositions are types, implications are functions, etc. If you have the right foundation, I think the stuff ends up being quite legible. Mathlib is very abstract which makes it harder to approach as a beginner, but you could imagine a sort of literate programming approach where we walk students through building their own personal Mathlib, refactoring it to use higher abstractions as they build it up, etc. In a sense this is what a math education is today, but with a kind of half-formal, half-natural language and with a bunch of implicit assumptions about what techniques the student really knows (since details are generally omitted) vs. letting them use whatever macros/tactics/theorems they've learned/created in other courses.
This could all be especially powerful if the objects you're working with have good Widgets[0][1] so you could visualize and interact with various intermediate expressions. I see tons of potential here. The Lean games[2] also show the potential for a bright future here that's kind of along the lines of "build your own library around topic X" (the NN game has been posted here a few times, but it's actually a framework with other games too!).
[0] https://lean-lang.org/lean4/doc/examples/widgets.lean.html
[1] https://github.com/leanprover-community/ProofWidgets4/blob/m...
I wanted to justify this with the “Roger Bannister Effect”. The thought is that we’re held back psychologically by the idea of the impossible. It takes one person to do it. And now everyone can do it, freed from the mind trap. But further reading shows we were incrementally approaching what Roger Bannister did first: the 4 minute mile. And the pause before that record was likely not psychological but physical with World War Two. [0] And this jives with the TFA when Mr. Wolfram writes about a quarter of a century not yielding a human interpretation of his computer’s output.
All I’m left with is my anecdotes. I had a math professor in college who assigned homework every class. Since it was his first time teaching, he came up with the questions live. I’d come to class red in the face after struggling with questions all night. Then the professor would sheepishly reveal some of his statements were false. That unknown sapped a lot of motivation. Dead ends felt more conclusive. Falsehood was an easy scapegoat.
[0] https://www.scienceofrunning.com/2017/05/the-roger-bannister...
I'm skeptical of an in-principle impossibility of understanding complex proofs. I think computers likely will have, or already do have, a capability for explaining proofs piecemeal, with explanations that bridge from the proof itself to familiar intuitions.
Needing to understanding by increasingly sophisticated bridge explanations may be us getting further removed from immediate understanding but I don't think it crosses a magical threshold or anything that fundamentally calls into question how operational/useful the proofs are.
There is a famous quote by Riemann: "If only I had the theorems! Then I should find the proofs easily enough. "
Once you have a proof, simplifying it should be much easier, even for computers.
There are proofs created by humans that cannot be undersood by humans or for many few. For example the Fermat Last's Theorem proof.
Where in the article is this discussed?
More in detail: Proofs are nothing more than strings, and checking the validity of a proof can be done mechanically (and efficiently), so we can just enumerate all valid proofs up to length N and pick out its conclusion.
Ok, but what about the theorems humans would actually care about?
But there are some theorems many people care about and they don't have any known short proofs (e.g. Fermat's Last Theorem).
A simple question, since • is nand, the theorem ((a•b)•c)•(a•((a•c)•a))=c, can be proved trivially computing the value of both sides for the 8 values of a,b,c. Also there are 2^(2^2)=16 logic functions with two variables so is trivial to verify that the theorem is valid iff • is nand. Perhaps the difficulty is proving the theorem using rules?, there must be something that I don't see (I just took a glimpse in a few seconds).
Automatic formal proofs can be useful when you are able to test [1] https://doi.org/10.1016/j.jpaa.2008.11.043 [2] https://ceur-ws.org/Vol-3455/short4.pdf
They also permit shockingly short proofs compared to the customary notation. Which, incidentally was also devised by Peirce. Peano freely acknowledges all he did is change some of the symbols to avoid confusion (Peirce used capital sigma and pi for existential and universal quantification).
I loved Peirce's essays, but have not tried to read his work on logic or semiotics.
That should be plenty to get you started! Digging through the references in those papers and the bibliography on Sowa's site will find you plenty more modern Peirce scholarship. I think Peirce would be pleased that his seemingly abstract approach to logic ended up inspiring one of the most pragmatically useful classes of software ever, the relational database.
[1] https://www.jfsowa.com/pubs/egtut.pdf
[2] https://www.jfsowa.com/pubs/eg2cg.pdf
I was surprised and intrigued to find Wolfram's name when I was reading background literature doing this blog post https://www.philipzucker.com/cody_sheffer/ where I translated my friend's Lean proof of the correspondence of Sheffer stroke to Boolean algebra in my python proof assistant knuckledragger https://github.com/philzook58/knuckledragger (an easier result than Wolfram's single axiom one).
I was also intrigued reading about the equational theorem prover Waldmeister https://www.mpi-inf.mpg.de/departments/automation-of-logic/s... that it may have been acquired by Mathematica? It's hard to know how Mathematica's equational reasoning system is powered behind the scenes.
Mathematica can really make for some stunning visuals. Quite impressive.
But then again, when "|" is NAND, it follows that (a|b)|(a|b) means NOT (a|b), which breaks the apparent link to symmetric spaces, because x reflected through itself should give back x.