Goto

Collaborating Authors

 debian


Discover and Mitigate Unknown Biases with Debiasing Alternate Networks

Li, Zhiheng, Hoogs, Anthony, Xu, Chenliang

arXiv.org Artificial Intelligence

Deep image classifiers have been found to learn biases from datasets. To mitigate the biases, most previous methods require labels of protected attributes (e.g., age, skin tone) as full-supervision, which has two limitations: 1) it is infeasible when the labels are unavailable; 2) they are incapable of mitigating unknown biases -- biases that humans do not preconceive. To resolve those problems, we propose Debiasing Alternate Networks (DebiAN), which comprises two networks -- a Discoverer and a Classifier. By training in an alternate manner, the discoverer tries to find multiple unknown biases of the classifier without any annotations of biases, and the classifier aims at unlearning the biases identified by the discoverer. While previous works evaluate debiasing results in terms of a single bias, we create Multi-Color MNIST dataset to better benchmark mitigation of multiple biases in a multi-bias setting, which not only reveals the problems in previous methods but also demonstrates the advantage of DebiAN in identifying and mitigating multiple biases simultaneously. We further conduct extensive experiments on real-world datasets, showing that the discoverer in DebiAN can identify unknown biases that may be hard to be found by humans. Regarding debiasing, DebiAN achieves strong bias mitigation performance.


Learning Embeddings of Directed Networks with Text-Associated Nodes---with Applications in Software Package Dependency Networks

Zhong, Shudan, Xu, Hong

arXiv.org Machine Learning

A network embedding consists of a vector representation for each node in the network. Network embeddings have shown their usefulness in node classification and visualization in many real-world application domains, such as social networks and web networks. While directed networks with text associated with each node, such as citation networks and software package dependency networks, are commonplace, to the best of our knowledge, their embeddings have not been specifically studied. In this paper, we create PCT ADW-1 and PCT ADW-2, two algorithms based on NNs that learn em-beddings of directed networks with text associated with each node. We create two new labeled directed networks with text-associated node: The package dependency networks in two popular GNU/Linux distributions, Debian and Fedora. We experimentally demonstrate that the embeddings produced by our NNs resulted in node classification with better quality than those of various baselines on these two networks. We observe that there exist systematic presence of analogies (similar to those in word embeddings) in the network embed-dings of software package dependency networks. To the best of our knowledge, this is the first time that such a systematic presence of analogies is observed in network and document embeddings. This may potentially open up a new venue for better understanding networks and documents algorith-mically using their embeddings as well as for better human understanding of network and document embeddings.


Choosing a Linix Operating System for ML • /r/MachineLearning

@machinelearnbot

Linux refers to the operating system's kernel. The different distributions are simply different collections of software, packaged in slightly different ways. In view of this, I'd be surprised if you find you can't run a certain ML framework on one Linux distro, if it runs on another. Having said that, to make things as easy as possible to get started, I've noticed that most projects tend to release fully packaged binaries for at least Debian, Ubuntu, and RedHat. Therefore it might be your best option to pick one of those.