Free Lunch for Pass@$k$? Low Cost Diverse Sampling for Diffusion Language Models
arXiv:2603.04893v1 Announce Type: new Abstract: Diverse outputs in text generation are necessary for effective exploration in complex reasoning tasks, such as code generation and mathematical problem solving. Such Pass@$k$ problems benefit from distinct candidates covering the solution space. However, traditional sampling approaches often waste computational resources on repetitive failure modes. While Diffusion Language Models have emerged as a competitive alternative to the prevailing Autoregressive paradigm, they remain susceptible to this redundancy, with independent samples frequently collapsing into similar modes. To address this, we propose a training free, low cost intervention to enhance generative diversity in Diffusion Language Models. Our approach modifies intermediate samples in a batch sequentially, where each sample is repelled from the feature space of previous samples, actively penalising redundancy. Unlike prior methods that require retraining or beam search, our str
arXiv:2603.04893v1 Announce Type: new Abstract: Diverse outputs in text generation are necessary for effective exploration in complex reasoning tasks, such as code generation and mathematical problem solving. Such Pass@$k$ problems benefit from distinct candidates covering the solution space. However, traditional sampling approaches often waste computational resources on repetitive failure modes. While Diffusion Language Models have emerged as a competitive alternative to the prevailing Autoregressive paradigm, they remain susceptible to this redundancy, with independent samples frequently collapsing into similar modes. To address this, we propose a training free, low cost intervention to enhance generative diversity in Diffusion Language Models. Our approach modifies intermediate samples in a batch sequentially, where each sample is repelled from the feature space of previous samples, actively penalising redundancy. Unlike prior methods that require retraining or beam search, our strategy incurs negligible computational overhead, while ensuring that each sample contributes a unique perspective to the batch. We evaluate our method on the HumanEval and GSM8K benchmarks using the LLaDA-8B-Instruct model. Our results demonstrate significantly improved diversity and Pass@$k$ performance across various temperature settings. As a simple modification to the sampling process, our method offers an immediate, low-cost improvement for current and future Diffusion Language Models in tasks that benefit from diverse solution search. We make our code available at https://github.com/sean-lamont/odd.
Executive Summary
The article proposes a low-cost intervention to enhance generative diversity in Diffusion Language Models. The approach modifies intermediate samples in a batch to penalize redundancy, ensuring each sample contributes a unique perspective. This method is training-free, incurs negligible computational overhead, and demonstrates improved diversity and Pass@$k$ performance across various temperature settings. The approach is evaluated on the HumanEval and GSM8K benchmarks using the LLaDA-8B-Instruct model, with significant improvements observed. The code is made available, offering an immediate, low-cost improvement for current and future Diffusion Language Models in tasks that benefit from diverse solution search.
Key Points
- ▸ Low-cost intervention to enhance generative diversity in Diffusion Language Models
- ▸ Training-free approach with negligible computational overhead
- ▸ Improved diversity and Pass@$k$ performance across various temperature settings
Merits
Efficient Sampling
The proposed method efficiently samples diverse outputs without wasting computational resources on repetitive failure modes.
Improved Performance
The approach demonstrates significant improvements in diversity and Pass@$k$ performance, making it a valuable contribution to the field.
Demerits
Limited Evaluation
The evaluation is limited to two benchmarks, and further testing on a broader range of tasks and models is necessary to fully assess the approach's effectiveness.
Expert Commentary
The article presents a significant contribution to the field of natural language processing, particularly in the context of Diffusion Language Models. The proposed approach addresses a critical limitation of traditional sampling methods, which often result in repetitive failure modes. By modifying intermediate samples to penalize redundancy, the method efficiently samples diverse outputs, demonstrating improved performance on benchmarks. While further evaluation is necessary, the approach has the potential to improve the effectiveness of language models in tasks that benefit from diverse solution search, such as code generation and mathematical problem solving.
Recommendations
- ✓ Further evaluation of the approach on a broader range of tasks and models to fully assess its effectiveness
- ✓ Exploration of potential applications of the method in other areas of AI research, such as computer vision and robotics