wandb
- Europe > France (0.14)
- North America > United States > California > Alameda County > Berkeley (0.04)
Efficient Rotation Invariance in Deep Neural Networks through Artificial Mental Rotation
Tuggener, Lukas, Stadelmann, Thilo, Schmidhuber, Jürgen
Humans and animals recognize objects irrespective of the beholder's point of view, which may drastically change their appearances. Artificial pattern recognizers also strive to achieve this, e.g., through translational invariance in convolutional neural networks (CNNs). However, both CNNs and vision transformers (ViTs) perform very poorly on rotated inputs. Here we present artificial mental rotation (AMR), a novel deep learning paradigm for dealing with in-plane rotations inspired by the neuro-psychological concept of mental rotation. Our simple AMR implementation works with all common CNN and ViT architectures. We test it on ImageNet, Stanford Cars, and Oxford Pet. With a top-1 error (averaged across datasets and architectures) of $0.743$, AMR outperforms the current state of the art (rotational data augmentation, average top-1 error of $0.626$) by $19\%$. We also easily transfer a trained AMR module to a downstream task to improve the performance of a pre-trained semantic segmentation model on rotated CoCo from $32.7$ to $55.2$ IoU.
- Europe > Switzerland (0.04)
- Asia > Japan > Honshū > Tōhoku > Fukushima Prefecture > Fukushima (0.04)
- Europe > Italy > Veneto > Venice (0.04)
- Asia > Middle East > Saudi Arabia (0.04)
Transformers for Multi-Regression -- [PART2] – Towards AI
Originally published on Towards AI the World's Leading AI and Technology News and Media Company. If you are building an AI-related product or service, we invite you to consider becoming an AI sponsor. At Towards AI, we help scale AI and technology startups. Let us help you unleash your technology to the masses. In the context of the FB3 competition, we aim to model six analysis metrics using pre-scored argumentative essays written by 8th-12th grade English Language Learners.
Machine Learning Experiment Tracking
At first glance, building and deploying machine learning models looks a lot like writing code. Tracking experiments in an organized way helps with all of these core issues. Weights and Biases (wandb) is a simple tool that helps individuals to track their experiments -- I talked to several machine learning leaders of different size teams about how they use wandb to track their experiments. The essential unit of progress in an ML project is an experiment, so most people track what they're doing somehow -- generally I see practitioners start with a spreadsheet or a text file to keep track of what they're doing. Spreadsheets and docs are incredibly flexible -- what's wrong with this approach?
GitHub - LeoGrin/tabular-benchmark
Accompanying repository for the paper Why do tree-based models still outperform deep learning on tabular data? To download these datasets, simply run python data/download_data.py. We're planning to release a version allowing to use Benchopt instead of WandB to make it easier to run. All the R code used to generate the analyses and figures in available in the analyses folder. The datasets used in the benchmark have been uploaded as OpenML benchmarks, with the same transformations that are used in the paper.
WandB: Alternative to TensorBoard more intresting thing about.
Wandb has many more the most best thing is that we never loose the procfiling over wandb but instread on tensorflow might we can loose . If you are model craetion idea working with Pytorch or Tensoflow here you good to go . But we have many more on wandb.rather Here I would be using MNIST dataset; this dataset consists of hand-written digits from 0 to 9 so our task would be to create a model that classifies these digits.
StoryDB: Broad Multi-language Narrative Dataset
Tikhonov, Alexey, Samenko, Igor, Yamshchikov, Ivan P.
This paper presents StoryDB - a broad multi-language dataset of narratives. StoryDB is a corpus of texts that includes stories in 42 different languages. Every language includes 500+ stories. Some of the languages include more than 20 000 stories. Every story is indexed across languages and labeled with tags such as a genre or a topic. The corpus shows rich topical and language variation and can serve as a resource for the study of the role of narrative in natural language processing across various languages including low resource ones. We also demonstrate how the dataset could be used to benchmark three modern multilanguage models, namely, mDistillBERT, mBERT, and XLM-RoBERTa.
- North America > United States > Colorado (0.04)
- Europe > Russia > Northwestern Federal District > Leningrad Oblast > Saint Petersburg (0.04)
- Europe > Germany > Berlin (0.04)
- Asia > Russia > Siberian Federal District > Novosibirsk Oblast > Novosibirsk (0.04)
Multi-label Emotion Classification with PyTorch + HuggingFace's Transformers and W&B for Tracking
The GoEmotions dataset contains 58k carefully curated Reddit comments labeled for 27 emotion categories or Neutral. The raw data is included as well as the smaller, simplified version of the dataset with predefined train/val/test splits. After going through a few examples in this dataset on their visualizer, I realized that this is an extremely crucial dataset because it's rare to find sentiment classifier datasets that go beyond 5–6 emotions. But here, we have 27 emotions being assigned, with rare and close enough emotions like disappointment, disapproval, grief, remorse, sadness, etc. Detecting such close enough emotions is often difficult in typical datasets. This made it clear to me that this is an excellent dataset that can be scaled for usage in many applications that involve text analysis.
EENLP: Cross-lingual Eastern European NLP Index
Tikhonov, Alexey, Malkhasov, Alex, Manoshin, Andrey, Dima, George, Cserháti, Réka, Asif, Md. Sadek Hossain, Sárdi, Matt
This report presents the results of the EENLP project, done as a part of EEML 2021 summer school. It presents a broad index of NLP resources for Eastern European languages, which, we hope, could be helpful for the NLP community; several new hand-crafted cross-lingual datasets focused on Eastern European languages, and a sketch evaluation of cross-lingual transfer learning abilities of several modern multilingual Transformer-based models.