← Back to Mathematics

Mathematics

AI and Machine Learning in Number Theory: Predicting Class Numbers of Real Quadratic Fields

  • AI in mathematics
  • machine learning in number theory
  • AutoML research
  • class number problem
  • real quadratic fields
  • Dedekind zeta function
  • symbolic classification
  • LightGBM
  • CatBoost
  • artificial intelligence research

MLJAR tools were used in the following publication.

Machine Learning Class Numbers of Real Quadratic Fields

Malik Amir, Yang-Hui He, Kyu-Hwan Lee, Thomas Oliver, Eldar Sultanow

Microsoft Research New England, USA | City, University of London, UK | Simons Foundation (Grant #712100) | EPSRC Research Grant EP/S032460/1

This research demonstrates how artificial intelligence and machine learning can be applied to fundamental problems in number theory, including the classification of real quadratic fields by their class numbers. Using LightGBM, CatBoost, symbolic classification, and automated feature engineering, the study shows how AI models can rediscover deep mathematical structures such as genus theory and the analytic class number formula. The work introduces a cost-function framework to quantify separability of arithmetic invariants and highlights the role of explainable AI in mathematical discovery. This case study illustrates how AutoML and data-driven methods can accelerate research in pure mathematics and computational number theory.

International Journal of Data Science in the Mathematical Sciences • September 19, 2022

DOI: https://arxiv.org/abs/2209.09283

Research Domains

Explore peer-reviewed and applied machine learning studies across diverse domains, including healthcare analytics, financial modeling, manufacturing optimization, and structured data classification problems.

Why Researchers and ML Engineers Choose MLJAR Studio

A private, AI-powered Python notebook designed for reproducible machine learning experiments, structured benchmarking, and applied research workflows - fully under your control.

Reproducible Machine Learning Experiments

Design structured pipelines, save experiment runs, and compare results across iterations with full transparency. Every validation setup, hyperparameter configuration, and model benchmark is recorded - making your research repeatable and defensible.

Local-First Execution & Data Control

Run all workflows directly on your machine. Sensitive datasets remain private, with no mandatory cloud uploads or external AI services required. Maintain full control over runtime environments and compliance requirements.

Autonomous Model Benchmarking & Optimization

Automatically compare candidate models, perform cross-validation, and run hyperparameter optimization while retaining full visibility into generated Python code and evaluation metrics. Accelerate experimentation without sacrificing methodological rigor.

Build Research-Grade ML Workflows Locally

Run automated model benchmarking, hyperparameter optimization, and autonomous experiments while keeping full control over your data.