We are a team at Yale and MIT who is building a new, fine-grained benchmark
to evaluate the correctness of LLM-generated mathematical proofs. We are
looking for domain experts to help us ensure these benchmarks are rigorous
and reproducible.
Who we are looking for:
We are recruiting Faculty, Researchers, and PhD Students (current
candidates or degree holders) with expertise in upper-level mathematics.
The Incentive:
We are offering paper co-authorship to evaluators who complete ~20 hours of
high-quality reviews before December 31, 2025.
The Role:
We need experts in Algebra, Number Theory, ODE, Analysis, Topology, Number
Theory, Combinatorics, and other upper-level fields to:
Review model-generated proofs (blind to identity).
Score them against a rubric.
Provide brief rationales for your scores.
Logistics:
Deadline: December 31st, 2025
Time Commitment: ~20 hours (Flexible/Asynchronous)
Remote
Apply Here: https://forms.gle/UcSGpGhwfKdvCS8E9
If you are interested in how we measure the true reasoning capabilities of
AI in mathematics, we'd love to have you on the team.
Questions? Email qql@mit.edu
**********************************************************
*
* Contributions to be spread via DMANET are submitted to
*
* DMANET@zpr.uni-koeln.de
*
* Replies to a message carried on DMANET should NOT be
* addressed to DMANET but to the original sender. The
* original sender, however, is invited to prepare an
* update of the replies received and to communicate it
* via DMANET.
*
* DISCRETE MATHEMATICS AND ALGORITHMS NETWORK (DMANET)
* http://www.zaik.uni-koeln.de/AFS/publications/dmanet/
*
**********************************************************