Goto

Collaborating Authors

 resolvability


Metric Dimension and Resolvability of Jaccard Spaces

Lladser, Manuel E., Paradise, Alexander J.

arXiv.org Artificial Intelligence

A subset of points in a metric space is said to resolve it if each point in the space is uniquely characterized by its distance to each point in the subset. In particular, resolving sets can be used to represent points in abstract metric spaces as Euclidean vectors. Importantly, due to the triangle inequality, points close by in the space are represented as vectors with similar coordinates, which may find applications in classification problems of symbolic objects under suitably chosen metrics. In this manuscript, we address the resolvability of Jaccard spaces, i.e., metric spaces of the form $(2^X,\text{Jac})$, where $2^X$ is the power set of a finite set $X$, and $\text{Jac}$ is the Jaccard distance between subsets of $X$. Specifically, for different $a,b\in 2^X$, $\text{Jac}(a,b)=|a\Delta b|/|a\cup b|$, where $|\cdot|$ denotes size (i.e., cardinality) and $\Delta$ denotes the symmetric difference of sets. We combine probabilistic and linear algebra arguments to construct highly likely but nearly optimal (i.e., of minimal size) resolving sets of $(2^X,\text{Jac})$. In particular, we show that the metric dimension of $(2^X,\text{Jac})$, i.e., the minimum size of a resolving set of this space, is $\Theta(|X|/\ln|X|)$. In addition, we show that a much smaller subset of $2^X$ suffices to resolve, with high probability, all different pairs of subsets of $X$ of cardinality at most $\sqrt{|X|}/\ln|X|$, up to a factor.


Can Error Mitigation Improve Trainability of Noisy Variational Quantum Algorithms?

Wang, Samson, Czarnik, Piotr, Arrasmith, Andrew, Cerezo, M., Cincio, Lukasz, Coles, Patrick J.

arXiv.org Artificial Intelligence

Variational Quantum Algorithms (VQAs) are often viewed as the best hope for near-term quantum advantage. However, recent studies have shown that noise can severely limit the trainability of VQAs, e.g., by exponentially flattening the cost landscape and suppressing the magnitudes of cost gradients. Error Mitigation (EM) shows promise in reducing the impact of noise on near-term devices. Thus, it is natural to ask whether EM can improve the trainability of VQAs. In this work, we first show that, for a broad class of EM strategies, exponential cost concentration cannot be resolved without committing exponential resources elsewhere. This class of strategies includes as special cases Zero Noise Extrapolation, Virtual Distillation, Probabilistic Error Cancellation, and Clifford Data Regression. Second, we perform analytical and numerical analysis of these EM protocols, and we find that some of them (e.g., Virtual Distillation) can make it harder to resolve cost function values compared to running no EM at all. As a positive result, we do find numerical evidence that Clifford Data Regression (CDR) can aid the training process in certain settings where cost concentration is not too severe. Our results show that care should be taken in applying EM protocols as they can either worsen or not improve trainability. On the other hand, our positive results for CDR highlight the possibility of engineering error mitigation methods to improve trainability.


An impossibility theorem concerning positive involvement in voting

Holliday, Wesley H.

arXiv.org Artificial Intelligence

In social choice theory with ordinal preferences, a voting method satisfies the axiom of positive involvement if adding to a preference profile a voter who ranks an alternative uniquely first cannot cause that alternative to go from winning to losing. In this note, we prove a new impossibility theorem concerning this axiom: there is no ordinal voting method satisfying positive involvement that also satisfies the Condorcet winner and loser criteria, resolvability, and a common invariance property for Condorcet methods, namely that the choice of winners depends only on the ordering of majority margins by size.


Split Cycle: A New Condorcet Consistent Voting Method Independent of Clones and Immune to Spoilers

Holliday, Wesley H., Pacuit, Eric

arXiv.org Artificial Intelligence

A voting method is Condorcet consistent if in any election in which one candidate is preferred by majorities to each of the other candidates, this candidate--the Condorcet winner--is the unique winner of the election. Condorcet consistent voting methods form an important class of methods in the theory of voting (see, e.g., Fishburn 1977; Brams and Fishburn 2002, 8; Zwicker 2016, 2.4; Pacuit 2019, 3.1.1). Although Condorcet methods are not currently used in government elections, they have been used by several private organizations (see Wikipedia contributors 2020b) and in over 30,000 polls through the Condorcet Internet Voting Service (https://civs.cs.cornell.edu). Recent initiatives in the U.S. to make available Instant Runoff Voting (Kambhampaty 2019), which uses the same ranked ballots needed for Condorcet methods, bring Condorcet methods closer to political application. Indeed, Eric Maskin and Amartya Sen have recently proposed the use of Condorcet methods in U.S. presidential primaries (Maskin and Sen 2016, 2017a,b). In the meantime, Condorcet methods continue to be used by committees, clubs, etc.


Finite-sample risk bounds for maximum likelihood estimation with arbitrary penalties

Brinda, W. D., Klusowski, Jason M.

arXiv.org Machine Learning

Remarkably general method for bounding the statistical risk of penalized likelihood estimators comes from work on two-part coding, one of the minimum description length (MDL) approaches to statistical inference. Two-part coding MDL prescribes assigning codelengths to a model (or model class) then selecting the distribution that provides the most efficient description of one's data [1]. The total description length has two parts: the part that specifies a distribution within the model (as well as a model within the model class if necessary) and the part that specifies the data with reference to the specified distribution. If the codelengths are exactly Kraft-valid, this approach is equivalent to Bayesian maximum a posteriori (MAP) estimation, in that the two parts correspond to log reciprocal of prior and log reciprocal of likelihood respectively. More generally, one can call the part of the codelength specifying the distribution a penalty term; it is called the complexity in MDL literature. Let (Θ, L) denote a discrete set indexing distributions along with a complexity function. With X P, the (pointwise) redundancy of any θ Θ is its two-part codelength minus log(1/p(X)), the codelength one gets by using P as the coding distribution.