Goto

Collaborating Authors

 maximizing


MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

Neural Information Processing Systems

The strong correlation between neurons or filters can significantly weaken the generalization ability of neural networks. Inspired by the well-known Tammes problem, we propose a novel diversity regularization method to address this issue, which makes the normalized weight vectors of neurons or filters distributed on a hypersphere as uniformly as possible, through maximizing the minimal pairwise angles (MMA). This method can easily exert its effect by plugging the MMA regularization term into the loss function with negligible computational overhead. The MMA regularization is simple, efficient, and effective. Therefore, it can be used as a basic regularization method in neural network training. Extensive experiments demonstrate that MMA regularization is able to enhance the generalization ability of various modern models and achieves considerable performance improvements on CIFAR100 and TinyImageNet datasets. In addition, experiments on face verification show that MMA regularization is also effective for feature learning.


Maximizing the efficiency of human feedback in AI alignment: a comparative analysis

Chouliaras, Andreas, Chatzopoulos, Dimitris

arXiv.org Artificial Intelligence

Reinforcement Learning from Human Feedback (RLHF) relies on preference modeling to align machine learning systems with human values, yet the popular approach of random pair sampling with Bradley-Terry modeling is statistically limited and inefficient under constrained annotation budgets. In this work, we explore alternative sampling and evaluation strategies for preference inference in RLHF, drawing inspiration from areas such as game theory, statistics, and social choice theory. Our best-performing method, Swiss InfoGain, employs a Swiss tournament system with a proxy mutual-information-gain pairing rule, which significantly outperforms all other methods in constrained annotation budgets while also being more sample-efficient. Even in high-resource settings, we can identify superior alternatives to the Bradley-Terry baseline. Our experiments demonstrate that adaptive, resource-aware strategies reduce redundancy, enhance robustness, and yield statistically significant improvements in preference learning, highlighting the importance of balancing alignment quality with human workload in RLHF pipelines.


Review for NeurIPS paper: The Adaptive Complexity of Maximizing a Gross Substitutes Valuation

Neural Information Processing Systems

Strengths: Soundness of the claims: The authors fully justify their claim. While most of the proofs are in the appendix, authors provide a high level sketch with intuition in the main body of the paper. Significance and novelty of the contribution: Authors provide best adaptive algorithms for maximizing a gross-substitutes function subject to cardinality constraint. Gross-substitutes is an important class of set functions. The authors' results show that they have obtained the best bound possible.


Review for NeurIPS paper: MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

Neural Information Processing Systems

Weaknesses: The regularization coefficient is important. It can be seen from Figure 3 that the influences of different coefficients are unstable in a single experiment. To better clarify the influences, the author should report the average results of multiple experiments under the same setting. I wonder if the convergence speed and stability are changed after introducing the MMA regularization in the training process. Therefore, it would be better to give the training loss curves with and without MMA regularization.


Review for NeurIPS paper: MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

Neural Information Processing Systems

The paper has initially received mixed reviews, but post-rebuttal the expert reviewers have converged to the decision that the paper is above the acceptance threshold and that the proposed regularization is of wide interest to the community. The authors are encouraged to incorporate the extra experimental results from the rebuttal into the final version of the paper. Also, the related work section should be revised by incorporating relevant works pointed by the reviewers (and as promised in the rebuttal).


Review for NeurIPS paper: Towards Maximizing the Representation Gap between In-Domain & Out-of-Distribution Examples

Neural Information Processing Systems

Weaknesses: Given that the major contribution of the paper is to improve detection of misclassified and OOD samples, it is a bit disappointing that the misclassification results in Table 2 and 8 are comparable at best with competing methods. In Fig 5, I'm curious what is the transform applied to the y-axis? L518 of the Appendix says: "In practice, since we do not know the characteristics of the OOD test examples, it may not be suitable to use a binary classifier for OOD detection tasks." This valid criticism also applies to the proposed method, since examples of outliers are used. Using binary classifiers has been shown to be very effective for highly practical tasks, for example, [3].


Rank It, Then Ask It: Input Reranking for Maximizing the Performance of LLMs on Symmetric Tasks

Dehghankar, Mohsen, Asudeh, Abolfazl

arXiv.org Artificial Intelligence

Large language models (LLMs) have quickly emerged as practical and versatile tools that provide new solutions for a wide range of domains. In this paper, we consider the application of LLMs on symmetric tasks where a query is asked on an (unordered) bag of elements. Examples of such tasks include answering aggregate queries on a database table. In general, when the bag contains a large number of elements, LLMs tend to overlook some elements, leading to challenges in generating accurate responses to the query. LLMs receive their inputs as ordered sequences. However, in this problem, we leverage the fact that the symmetric input is not ordered, and reordering should not affect the LLM's response. Observing that LLMs are less likely to miss elements at certain positions of the input, we introduce the problem of LLM input reranking: to find a ranking of the input that maximizes the LLM's accuracy for the given query without making explicit assumptions about the query. Finding the optimal ranking requires identifying (i) the relevance of each input element for answering the query and (ii) the importance of each rank position for the LLM's attention. We develop algorithms for estimating these values efficiently utilizing a helper LLM. We conduct comprehensive experiments on different synthetic and real datasets to validate our proposal and to evaluate the effectiveness of our proposed algorithms. Our experiments confirm that our reranking approach improves the accuracy of the LLMs on symmetric tasks by up to $99\%$ proximity to the optimum upper bound.


MMA Regularization: Decorrelating Weights of Neural Networks by Maximizing the Minimal Angles

Neural Information Processing Systems

The strong correlation between neurons or filters can significantly weaken the generalization ability of neural networks. Inspired by the well-known Tammes problem, we propose a novel diversity regularization method to address this issue, which makes the normalized weight vectors of neurons or filters distributed on a hypersphere as uniformly as possible, through maximizing the minimal pairwise angles (MMA). This method can easily exert its effect by plugging the MMA regularization term into the loss function with negligible computational overhead. The MMA regularization is simple, efficient, and effective. Therefore, it can be used as a basic regularization method in neural network training.


New Directions in Text Classification Research: Maximizing The Performance of Sentiment Classification from Limited Data

Agustian, Surya, Syah, Muhammad Irfan, Fatiara, Nurul, Abdillah, Rahmad

arXiv.org Artificial Intelligence

The stakeholders' needs in sentiment analysis for various issues, whether positive or negative, are speed and accuracy. One new challenge in sentiment analysis tasks is the limited training data, which often leads to suboptimal machine learning models and poor performance on test data. This paper discusses the problem of text classification based on limited training data (300 to 600 samples) into three classes: positive, negative, and neutral. A benchmark dataset is provided for training and testing data on the issue of Kaesang Pangarep's appointment as Chairman of PSI. External data for aggregation and augmentation purposes are provided, consisting of two datasets: the topic of Covid Vaccination sentiment and an open topic. The official score used is the F1-score, which balances precision and recall among the three classes, positive, negative, and neutral. A baseline score is provided as a reference for researchers for unoptimized classification methods. The optimized score is provided as a reference for the target score to be achieved by any proposed method. Both scoring (baseline and optimized) use the SVM method, which is widely reported as the state-of-the-art in conventional machine learning methods. The F1-scores achieved by the baseline and optimized methods are 40.83% and 51.28%, respectively.


Boosting Algorithms for Maximizing the Soft Margin

Neural Information Processing Systems

We present a novel boosting algorithm, called SoftBoost, designed for sets of bi- nary labeled examples that are not necessarily separable by convex combinations of base hypotheses. Our algorithm achieves robustness by capping the distribu- tions on the examples. Our update of the distribution is motivated by minimizing a relative entropy subject to the capping constraints and constraints on the edges of the obtained base hypotheses. The capping constraints imply a soft margin in the dual optimization problem. Our algorithm produces a convex combination of hypotheses whose soft margin is within δ of its maximum.