Friday, December 12, 2025

[DMANET] Seeking LLM Math Proof Evaluators

Call for Mathematicians: Help Evaluate LLM Math Proofs

We are a team at Yale and MIT who is building a new, fine-grained benchmark
to evaluate the correctness of LLM-generated mathematical proofs. We are
looking for domain experts to help us ensure these benchmarks are rigorous
and reproducible.

Who we are looking for:

We are recruiting Faculty, Researchers, and PhD Students (current
candidates or degree holders) with expertise in upper-level mathematics.

The Incentive:

We are offering paper co-authorship to evaluators who complete ~20 hours of
high-quality reviews before December 31, 2025.

The Role:

We need experts in Algebra, Number Theory, ODE, Analysis, Topology, Number
Theory, Combinatorics, and other upper-level fields to:

Review model-generated proofs (blind to identity).
Score them against a rubric.
Provide brief rationales for your scores.

Logistics:

Deadline: December 31st, 2025
Time Commitment: ~20 hours (Flexible/Asynchronous)
Remote

Apply Here: https://forms.gle/UcSGpGhwfKdvCS8E9

If you are interested in how we measure the true reasoning capabilities of
AI in mathematics, we'd love to have you on the team.

Questions? Email qql@mit.edu

**********************************************************
*
* Contributions to be spread via DMANET are submitted to
*
* DMANET@zpr.uni-koeln.de
*
* Replies to a message carried on DMANET should NOT be
* addressed to DMANET but to the original sender. The
* original sender, however, is invited to prepare an
* update of the replies received and to communicate it
* via DMANET.
*
* DISCRETE MATHEMATICS AND ALGORITHMS NETWORK (DMANET)
* http://www.zaik.uni-koeln.de/AFS/publications/dmanet/
*
**********************************************************