Goto

Collaborating Authors

 Chen, Xinjia


Learn to Cluster Faces with Better Subgraphs

arXiv.org Artificial Intelligence

Face clustering can provide pseudo-labels to the massive unlabeled face data and improve the performance of different face recognition models. The existing clustering methods generally aggregate the features within subgraphs that are often implemented based on a uniform threshold or a learned cutoff position. This may reduce the recall of subgraphs and hence degrade the clustering performance. This work proposed an efficient neighborhood-aware subgraph adjustment method that can significantly reduce the noise and improve the recall of the subgraphs, and hence can drive the distant nodes to converge towards the same centers. More specifically, the proposed method consists of two components, i.e. face embeddings enhancement using the embeddings from neighbors, and enclosed subgraph construction of node pairs for structural information extraction. The embeddings are combined to predict the linkage probabilities for all node pairs to replace the cosine similarities to produce new subgraphs that can be further used for aggregation of GCNs or other clustering methods. The proposed method is validated through extensive experiments against a range of clustering solutions using three benchmark datasets and numerical results confirm that it outperforms the SOTA solutions in terms of generalization capability.


Probability Estimation with Truncated Inverse Binomial Sampling

arXiv.org Machine Learning

In science and engineering, it is an ubiquitous problem to estimate the probability of event based on Monte Carlo simulation. For instance, in engineering technology, a critical c oncern is the probability of failure or risk, which is generally considered as the probability that certain pre -specified requirements for the relevant system are violated in the presence of uncertainties. Ever since th e advent of modern computers, extensive research works have been devoted to quantitative approaches o f risk evaluation for engineering systems (see, e.g., [1, 8, 9, 11, 16, 18, 20] and the references therein). I n additional to theoretical development, many softwares have been developed for risk evaluation. For exam ple, for control systems, a software called RACT has been developed for evaluating the risk of uncertain syste ms [7, 21]. Many softwares such as APMC [13], PRISM [15], UPPAAL [6], have been developed for evaluating t he risk of stochastic discrete event systems (see, [1] and the references therein). One of the remarkable achievements of existing theories and softw ares is the rigorous control of error in the estimation of probability, that is, the probability of relevant ev ent can be evaluated with certified reliability. Theoretically, for a priori given α, δ (0, 1), existing methods are able to produce an estimate null p for the true value of the probability p so that one can be 100(1 δ)% confident that null p p α holds. 1 Unfortunately, existing methods suffer from huge computational complexity as the margin of absolute error α is small, e.g. 10