cat
0e0157ce5ea15831072be4744cbd5334-Supplemental-Conference.pdf
Ep denotes the total number of epochs needed to fine-tunemodeloverthedataset. Consequently,itcan be minimized using asecond-order Newtonmethod. Wecan also detect some qualitativedifferences inthe attention maps atdifferent resolutions: The entropy, i.e. how much the attention is concentrated or spread across different tokens, changes significantlybetweenlevels. Hence, the similarity between consecutive representations is expected to be strong. On the other hand, when only looking atCascadeXML'spoints (inblue) inFigure 4,weobservethat thetasks in the first meta-classifier and the extreme classifier are substantially different. As shown in Table 11, the shortlisting achieves very good recall rates.
Cross Aggregation Transformer for Image Restoration
Recently, Transformer architecture has been introduced into image restoration to replace convolution neural network (CNN) with surprising results. Considering the high computational complexity of Transformer with global attention, some methods use the local square window to limit the scope of self-attention. However, these methods lack direct interaction among different windows, which limits the establishment of long-range dependencies. To address the above issue, we propose a new image restoration model, Cross Aggregation Transformer (CAT). The core of our CAT is the Rectangle-Window Self-Attention (Rwin-SA), which utilizes horizontal and vertical rectangle window attention in different heads parallelly to expand the attention area and aggregate the features cross different windows. We also introduce the Axial-Shift operation for different window interactions. Furthermore, we propose the Locality Complementary Module to complement the self-attention mechanism, which incorporates the inductive bias of CNN (e.g., translation invariance and locality) into Transformer, enabling global-local coupling. Extensive experiments demonstrate that our CAT outperforms recent state-of-the-art methods on several image restoration applications.
CATs: Cost Aggregation Transformers for Visual Correspondence
We propose a novel cost aggregation network, called Cost Aggregation Transformers (CATs), to find dense correspondences between semantically similar images with additional challenges posed by large intra-class appearance and geometric variations. Cost aggregation is a highly important process in matching tasks, which the matching accuracy depends on the quality of its output. Compared to hand-crafted or CNN-based methods addressing the cost aggregation, in that either lacks robustness to severe deformations or inherit the limitation of CNNs that fail to discriminate incorrect matches due to limited receptive fields, CATs explore global consensus among initial correlation map with the help of some architectural designs that allow us to fully leverage self-attention mechanism. Specifically, we include appearance affinity modeling to aid the cost aggregation process in order to disambiguate the noisy initial correlation maps and propose multi-level aggregation to efficiently capture different semantics from hierarchical feature representations. We then combine with swapping self-attention technique and residual connections not only to enforce consistent matching, but also to ease the learning process, which we find that these result in an apparent performance boost. We conduct experiments to demonstrate the effectiveness of the proposed model over the latest methods and provide extensive ablation studies.
Learning Conjoint Attentions for Graph Neural Nets
Besides considering the layer-wise node features propagated within the GNN, CAs can additionally incorporate various structural interventions, such as node cluster embedding, and higher-order structural correlations that can be learned outside of GNN, when computing attention scores. The node features that are regarded as significant by the conjoint criteria are therefore more likely to be propagated in the GNN. Given the novel Conjoint Attention strategies, we then propose Graph conjoint attention networks (CATs) that can learn representations embedded with significant latent features deemed by the Conjoint Attentions.
Computerized Adaptive Testing via Collaborative Ranking
As the deep integration of machine learning and intelligent education, Computerized Adaptive Testing (CAT) has received more and more research attention. Compared to traditional paper-and-pencil tests, CAT can deliver both personalized and interactive assessments by automatically adjusting testing questions according to the performance of students during the test process. Therefore, CAT has been recognized as an efficient testing methodology capable of accurately estimating a student's ability with a minimal number of questions, leading to its widespread adoption in mainstream selective exams such as the GMAT and GRE. However, just improving the accuracy of ability estimation is far from satisfactory in the real-world scenarios, since an accurate ranking of students is usually more important (e.g., in high-stakes exams). Considering the shortage of existing CAT solutions in student ranking, this paper emphasizes the importance of aligning test outcomes (student ranks) with the true underlying abilities of students. Along this line, different from the conventional independent testing paradigm among students, we propose a novel collaborative framework, Collaborative Computerized Adaptive Testing (CCAT), that leverages inter-student information to enhance student ranking.
Robot vacuums 'could water plants or play with cat'
The global household robots market size was valued at 10.3bn ( 7.7bn) in 2023 and is anticipated to hit 24.5bn by 2028, meaning such devices are an increasingly common sight in people's homes. Anyone who has watched a robot vacuum cleaner in action may argue these ideas are a little far-fetched, given that current machines sometimes struggle with the challenges presented by rugs and shoelaces while carrying out their core function. However, scientists from the University of Bath and the University of Calgary in Canada, have set out to prove that cleaners - and similar devices, such as lawnmowers - could be reprogrammed and modified relatively easily. Their study identified 100 functions the robots could possibly perform with simple adjustments. Other proposed tasks suggested by the scientists include a reprogrammed robot that carried the groceries from the car to the kitchen.
When I'm Sad My Computer Sends Me Cats
I wrote a program that sends cats to my phone when I'm sad at the computer. I was inspired by a tweet I saw last week. I've lost the link but, to paraphrase, it went something like this: I'm okay submitting myself to The Algorithm as long as it knows when I'm sad and forwards cats directly to my face I figured that you could probably solve this problem locally without leaking any personal data. Our computers are fast enough that we can run machine learning models in a browser in the background, maybe without even noticing. I went with vladmandic/human -- another strong contender was justadudewhohacks/face-api.js.
Edge-featured Graph Neural Architecture Search
Cai, Shaofei, Li, Liang, Han, Xinzhe, Zha, Zheng-jun, Huang, Qingming
Graph neural networks (GNNs) have been successfully applied to learning representation on graphs in many relational tasks. Recently, researchers study neural architecture search (NAS) to reduce the dependence of human expertise and explore better GNN architectures, but they over-emphasize entity features and ignore latent relation information concealed in the edges. To solve this problem, we incorporate edge features into graph search space and propose Edge-featured Graph Neural Architecture Search (EGNAS) to find the optimal GNN architecture. Specifically, we design rich entity and edge updating operations to learn high-order representations, which convey more generic message passing mechanisms. Moreover, the architecture topology in our search space allows to explore complex feature dependence of both entities and edges, which can be efficiently optimized by differentiable search strategy. Experiments at three graph tasks on six datasets show EGNAS can search better GNNs with higher performance than current state-of-the-art human-designed and searched-based GNNs.
Inference in Bayesian Networks
A Bayesian network is a compact, expressive representation of uncertain relationships among parameters in a domain. In this article, I introduce basic methods for computing with Bayesian networks, starting with the simple idea of summing the probabilities of events of interest. The article introduces major current methods for exact computation, briefly surveys approximation methods, and closes with a brief discussion of open issues. Often, truth is more elusive, and categorical statements can only be made by judgment of the likelihood or other ordinal attribute of competing propositions. Probability theory is the oldest and best-understood theory for representing and reasoning about such situations, but early AI experimental efforts at applying probability theory were disappointing and only confirmed a belief among AI researchers that those who worried about numbers were "missing the point."