โ† Back to Paper List

Is deep learning a useful tool for the pure mathematician?

Geordie Williamson
University of Sydney
arXiv (2023)
Reasoning Benchmark

๐Ÿ“ Paper Summary

AI for Mathematics Machine Learning for Combinatorics Conjecture Generation
Deep learning serves as a 'bicycle for the mind' in pure mathematics by identifying patterns in high-dimensional objects to guide conjecture formulation and counter-example search.
Core Problem
Pure mathematics problems often involve high-dimensional structures or complex invariants where traditional intuition fails, yet rigorous reasoning remains unsolved by current AI.
Why it matters:
  • Pure mathematicians often lack tools to explore high-dimensional spaces where intuition breaks down, limiting hypothesis generation
  • Current AI reasoning capabilities are insufficient for proving theorems, but pattern recognition capabilities are underutilized for guiding human proof attempts
  • Traditional brute-force search for counter-examples is computationally intractable for large combinatorial spaces
Concrete Example: When studying permutations in the symmetric group, the 'right descent set' is linearly predictable, but the 'left descent set' is hard to learn from vector inputs. A mathematician manually inspecting data might miss that simply changing the input representation to permutation matrices makes both sets equally learnable, revealing structural symmetries.
Key Novelty
Saliency-Guided Conjecture Generation
  • Train a neural network to predict a mathematical invariant (e.g., Kazhdan-Lusztig polynomial) from a structural input (e.g., Bruhat graph)
  • Use saliency analysis (gradients of the learned function) to identify which parts of the input structure most influence the output
  • Mathematicians inspect these high-saliency sub-structures to formulate precise conjectures about the underlying mathematical relationship
Evaluation Highlights
  • Graph Neural Network achieved ~98% accuracy predicting Kazhdan-Lusztig polynomials from Bruhat graphs (trained on ~20,000 graphs)
  • Neural network-guided search found a counter-example to a graph theory conjecture (regarding eigenvalue ฮป and matching number ยต) on a 19-vertex graph
  • Transformer model trained on matrices with Laplace-distributed eigenvalues generalized to positive-only eigenvalue matrices despite never seeing them during training
Breakthrough Assessment
7/10
While not proposing a new ML architecture, it successfully demonstrates a novel *workflow* (ML-guided intuition) that led to high-profile results in Nature, validating AI's utility in abstract math.
×