Retrospective on PAT x ICML 2026 AI Paper Assistant Program
March 30 2026 Retrospective on PAT x ICML 2026 AI Paper Assistant Program Gautam Kamath ICML 2026 By Google Researchers Rajesh Jayaram and Vincent Cohen-Addad, and ICML 2026 Program Chairs Alekh Agarwal, Miroslav Dudik, Sharon Li, Martin Jaggi. Following the success of the STOC Experimental AI Feedback program , we launched a similar experiment using Google Research’s Paper Assistant Tool (PAT) before the ICML 2026 deadline. Our goal was to provide authors with actionable, substantive feedback prior to the review process, allowing them to improve their submissions before the deadline. The PAT feedback was totally separate from the peer review process, and was not visible to reviewers, area chairs, program chairs, or anyone outside of the authors. Now that the program has concluded and the feedback from our author survey is in, we want to share some of the results. The program ran from January 14th to January 26th , providing feedback for approximately 4,500 papers. Papers sent to the system received feedback within ~30 minutes on average, with some edge cases (primarily due to large PDFs causing failures in the AI models). While we initially set conservative eligibility criteria to ensure a smooth rollout, we were able to quickly relax these requirements. Ultimately we offered a voucher to any author with an OpenReview account older than one month. We believe this afforded a maximal and equitable offering to the community, without allowing for misuse of the tool. We were quite pleased that we could make the program as accessible as possible. Highlights After concluding the program, we released a feedback survey to participating authors. We received 869 responses over the week it was open, and the results demonstrate a high level of satisfaction with the tool. In particular, the tool was able to frequently identify issues, and make suggestions which allowed the authors to make significant changes to their papers before the submission deadline. High Retention & Satisfaction: 92.1% of respondents stated they would use the tool again. Furthermore, 73.3% rated the feedback as “Very” or “Mostly” helpful. Only 1.6% found the tool to not be useful at all. Delivering Deep Feedback: The most encouraging results concern the depth of the feedback. The tool successfully identified suggestions beyond surface-level edits and typos: 35.4% of authors of papers containing theory reported the tool identified significant theory gaps that took more than an hour to fix. 31% of authors of papers with experimental results said the feedback prompted them to run new experiments. Clarity and Education: 87.3% of authors felt the tool improved the clarity and readability of their papers. Additionally, 84.5% saw clear educational value in the tool. Author Quotes on PAT: The PAT feedback was invaluable in improving my paper’s technical rigor before submission. It helped me identify and fix critical issues including: – Mathematical contradictions in objective functions – Inconsistencies in formal problem definitions – Unsupported claims in experimental sections – Missing theoretical justifications The structured, segment-by-segment analysis made it possible to address issues systematically under tight deadline pressure. — Apoorv Varshney, Department of Mathematics and Computing, Dr. B. R. Ambedkar National Institute of Technology Jalandhar Compared to most LLMs, PAT’s performance is already quite excellent. PAT discovered that one of my theorem proofs was not rigorous and provided some possible solutions. — Peihan Wu, Shanghai Normal University The system was very helpful and also rather “objective”. I look forward to seeing the “human” revisions and comparing them. — Paolo Stefano Giudici, Professor of Statistics, University of Pavia [PAT] gives a comprehensive review, allowing authors to address issues that could potentially be questioned by reviewers. — Enver Menadjiev, M.S.+Ph.D. student, Statistical AI Lab (SAILab) at Kim Jaechul Graduate School of AI at KAIST Outside feedback is so important when you’re about to submit. I [iterated] for two rounds with [competitor] before, which I paid for. PAT had deeper feedback. — Domenic Rosati, Dalhousie University How it worked The pipeline used for this experiment was scaffolding built on top of state-of-the-art Gemini-based models. To handle the complexity of technical papers, PAT segments the document into logical categorical sections, giving separate feedback for each section, with a high level summary at the beginning. Looking Ahead The PAT x ICML 2026 experiment shows that AI feedback can drive improvements in scientific work, enabling authors to fix theoretical gaps and conduct new experiments in advance of an expert-led peer review process. With the continued exponential growth in submissions to machine learning conferences, AI driven feedback has the benefit both of improving the quality of authors’ papers, and therefore the likelihood of acceptance and reproducibility, as well as easing the increasing burden on human reviewers. The PAT pipeline is still an experimental program in development. We plan to incorporate learnings from the ICML collaboration as we prepare for the next iteration of the Paper Assistant Tool. Thank you to the thousands of authors who participated in the program, provided feedback, and helped shape the future of AI-assisted research. We look forward to our continued collaboration to make our conference even better!
Executive Summary
The PAT x ICML 2026 AI Paper Assistant Program represents a pioneering experiment in leveraging AI-driven feedback to enhance academic paper submissions before peer review. Over two weeks, the program processed 4,500 papers, delivering actionable feedback within 30 minutes on average. Survey results (869 responses) revealed high satisfaction (92.1% retention rate, 73.3% found feedback helpful), with particularly strong performance in identifying theoretical gaps (35.4% for theory papers) and prompting additional experiments (31% for experimental papers). The tool also improved clarity (87.3%) and educational value (84.5%), demonstrating significant potential to democratize high-quality feedback in academic publishing. However, challenges such as handling large PDFs and ensuring equitable access remain.
Key Points
- ▸ AI-driven pre-submission feedback tool (PAT) deployed for ICML 2026 with rapid turnaround (avg. 30 mins).
- ▸ High author satisfaction (92.1% would reuse) and substantive feedback utility (e.g., 35.4% identified theory gaps).
- ▸ System operated independently of peer review, ensuring no contamination of the submission process.
Merits
Innovation in Academic Publishing
Pioneers scalable, AI-assisted pre-review feedback mechanisms, addressing a critical gap in academic publishing where authors often lack timely, high-quality feedback.
Equitable Access
Relaxed eligibility criteria to maximize participation, ensuring broad accessibility without misuse via OpenReview account verification.
Substantive Impact
Demonstrated ability to identify non-trivial issues (e.g., theoretical gaps, experimental design flaws) and improve paper quality across dimensions.
Educational Value
Enhanced clarity and readability for 87.3% of authors, serving as a de facto educational tool for researchers.
Demerits
Technical Limitations
Large PDFs caused processing failures in some edge cases, indicating scalability challenges in handling complex document formats.
Dependence on AI Quality
Feedback reliability hinges on the underlying AI model's accuracy, which may vary across disciplines (e.g., theoretical vs. experimental papers).
Survey Bias
Voluntary survey responses may not fully represent the entire participant pool, potentially skewing satisfaction metrics.
Expert Commentary
The PAT x ICML 2026 initiative marks a significant inflection point in academic publishing, where AI transcends mere automation to deliver substantive, actionable feedback. The program’s success—evidenced by high retention and deep technical insights—underscores the potential of AI to augment, rather than replace, human expertise. However, its limitations (e.g., PDF processing) and dependence on AI model quality necessitate caution. Future iterations should prioritize robustness across document formats and disciplinary diversity. The educational dimension of PAT (84.5% perceived value) is particularly noteworthy, suggesting a broader role for AI in researcher development. This experiment also raises critical questions about the evolving role of peer review: Could AI feedback reduce reviewer load or, conversely, introduce new biases? Policymakers and conference organizers must grapple with these trade-offs to harness AI’s benefits while preserving academic integrity. The program’s scalability and equitable design set a commendable benchmark for future AI-driven initiatives in scholarly communication.
Recommendations
- ✓ Expand PAT’s capability to handle complex document formats (e.g., LaTeX, large PDFs) to ensure universal applicability across research domains.
- ✓ Conduct longitudinal studies to assess the long-term impact of AI feedback on paper quality and reviewer workload in subsequent peer review cycles.
- ✓ Develop domain-specific AI models (e.g., for theory-heavy vs. experimental papers) to address current disparities in feedback utility.
- ✓ Establish a governance framework for AI-assisted tools in academic publishing, including transparency reports on model performance and bias mitigation strategies.
- ✓ Explore hybrid models where AI feedback is integrated with human expert reviews to balance scalability and nuance.
Sources
Original: ICML