Well File:






26b58a41da329e0cbde0cbf956640a58-AuthorFeedback.pdf

Neural Information Processing Systems

We first would like to thank all reviewers for their reviews and constructive comments. We give more details on some discussion points below. In fact, the symmetry assumption in our definition of a ranking loss (" Items are equivalent a priori" in Definition 3) Thanks for this remark, we will add it. R1: "does a convex calibrated surrogate in a given dimension exist if and only if there is a squared loss that is R2: "the loss L takes a tuple (Y,pi) as input, where pi is a predicted ranking. Yet, in many practical ranking tasks, the supervision is not a complete ranking.


A Other related work

Neural Information Processing Systems

In addition to the work on noisy convex optimization, the current paper is also thematically related to works in learning theory and complexity where the goal is to reconstruct simple classes of functions under outlier noise. This includes work on reconstruction of low-degree polynomials [4, 14, 15]. In particular, [15] gave an efficient algorithm whose error tolerance matches the information theoretic limits. In addition, recently, [9] achieved similar algorithmic guarantees for functions which are sparse in the Fourier space. While similar in spirit, the model in these works differ from the current paper in one crucial way - namely, while we only put a bound on the volume of the outlier locations, they, in addition, assume that the outlier locations are also uniformly distributed in the domain. At a more technical level, the results in [4, 14, 15, 9] crucially rely on techniques originating from coding theory such as the Goldreich-Levin theorem [13] and the Berlekamp-Welch algorithm [6].




Metacognitive Capabilities of LLMs: An Exploration in Mathematical Problem Solving, Nan Rosemary Ke

Neural Information Processing Systems

Metacognitive knowledge refers to humans' intuitive knowledge of their own thinking and reasoning processes. Today's best LLMs clearly possess some reasoning processes. The paper gives evidence that they also have metacognitive knowledge, including ability to name skills and procedures to apply given a task. We explore this primarily in context of math reasoning, developing a prompt-guided interaction procedure to get a powerful LLM to assign sensible skill labels to math questions, followed by having it perform semantic clustering to obtain coarser families of skill labels. These coarse skill labels look interpretable to humans. To validate that these skill labels are meaningful and relevant to the LLM's reasoning processes we perform the following experiments.