Windows 11 Pro may be the most underrated PC gaming upgrade ever at 15
TL;DR: Windows 11 Pro is down to 14.97 through June 1, its lowest price to date (reg. It's usually an afterthought, and that might be a mistake. Whether you're on Windows 10 or Windows 11 Home, upgrading to Windows 11 Pro may be the difference between an ordinary and an extraordinary gaming experience. Windows 11 Pro introduces DirectX 12 Ultimate, delivering higher frame rates, improved ray tracing, and lower latency for a smoother gaming experience. If you want faster load times and better graphics, this is your chance to optimize your rig.
Mind the Gap Between Prototypes and Images in Cross-domain Finetuning
In cross-domain few-shot classification (CFC), recent works mainly focus on adapting a simple transformation head on top of a frozen pre-trained backbone with few labeled data to project embeddings into a task-specific metric space where classification can be performed by measuring similarities between image instance and prototype representations. Technically, an assumption implicitly adopted in such a framework is that the prototype and image instance embeddings share the same representation transformation. However, in this paper, we find that there naturally exists a gap, which resembles the modality gap, between the prototype and image instance embeddings extracted from the frozen pre-trained backbone, and simply applying the same transformation during the adaptation phase constrains exploring the optimal representations and shrinks the gap between prototype and image representations. To solve this problem, we propose a simple yet effective method, contrastive prototype-image adaptation (CoPA), to adapt different transformations respectively for prototypes and images similarly to CLIP by treating prototypes as text prompts. Extensive experiments on Meta-Dataset demonstrate that CoPA achieves the state-of-the-art performance more efficiently. Meanwhile, further analyses also indicate that CoPA can learn better representation clusters, enlarge the gap, and achieve minimal validation loss at the enlarged gap.
Attention over learned object embeddings enables complex visual reasoning
Neural networks have achieved success in a wide array of perceptual tasks but often fail at tasks involving both perception and higher-level reasoning. On these more challenging tasks, bespoke approaches (such as modular symbolic components, independent dynamics models or semantic parsers) targeted towards that specific type of task have typically performed better. The downside to these targeted approaches, however, is that they can be more brittle than general-purpose neural networks, requiring significant modification or even redesign according to the particular task at hand. Here, we propose a more general neural-network-based approach to dynamic visual reasoning problems that obtains state-of-the-art performance on three different domains, in each case outperforming bespoke modular approaches tailored specifically to the task. Our method relies on learned object-centric representations, self-attention and self-supervised dynamics learning, and all three elements together are required for strong performance to emerge. The success of this combination suggests that there may be no need to trade off flexibility for performance on problems involving spatio-temporal or causal-style reasoning. With the right soft biases and learning objectives in a neural network we may be able to attain the best of both worlds.
One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
While reinforcement learning algorithms can learn effective policies for complex tasks, these policies are often brittle to even minor task variations, especially when variations are not explicitly provided during training. One natural approach to this problem is to train agents with manually specified variation in the training task or environment. However, this may be infeasible in practical situations, either because making perturbations is not possible, or because it is unclear how to choose suitable perturbation strategies without sacrificing performance. The key insight of this work is that learning diverse behaviors for accomplishing a task can directly lead to behavior that generalizes to varying environments, without needing to perform explicit perturbations during training. By identifying multiple solutions for the task in a single environment during training, our approach can generalize to new situations by abandoning solutions that are no longer effective and adopting those that are. We theoretically characterize a robustness set of environments that arises from our algorithm and empirically find that our diversity-driven approach can extrapolate to various changes in the environment and task.
We also apply PGD-1 and PGD-2 w/wo FN to attack a standard WRN-34-10 model, and PGD-1 (s, m): Heuristically, the value range of s is based on the averaged logit norms of standard
We thank all the reviewers for their valuable comments. Below, we address the detailed comments of each reviewer. The margin range m is chosen to be around cos 30 0.15. Then other samples that are not well-learned will dynamically contribute more. Figure 1, Sec. 3.5, and other comments: Thank you for the suggestions.
SoftBank taps Mizuho, SMBC, JPMorgan to lead 15 billion loan
SoftBank Group's investments in AI will be financed through a loan in which Mizuho Bank, Sumitomo Mitsui Banking and JPMorgan Chase serve as lead underwriters -- a sign of the Japanese tech investor's ability to secure financing for its outsized ambitions. The one-year 15 billion bridge loan, one of the biggest borrowings SoftBank has pulled off to date, will be financed by 21 banks and includes 1.35 billion from Mizuho, 1.25 billion from SMBC and 1 billion from JPMorgan, according to people familiar with the matter. It also includes a combined 950 million from HSBC Holdings and Barclays and 850 million jointly from seven banks including Goldman Sachs Group Inc., MUFG Bank and Credit Agricole, said the people, who asked not to be named as the details of the financing remain private. Representatives of the banks also declined to comment.
The NetHack Learning Environment Heinrich Küttler + Alexander H. Miller + Roberta Raileanu
Progress in Reinforcement Learning (RL) algorithms goes hand-in-hand with the development of challenging environments that test the limits of current methods. While existing RL environments are either sufficiently complex or based on fast simulation, they are rarely both. Here, we present the NetHack Learning Environment (NLE), a scalable, procedurally generated, stochastic, rich, and challenging environment for RL research based on the popular single-player terminalbased roguelike game, NetHack. We argue that NetHack is sufficiently complex to drive long-term research on problems such as exploration, planning, skill acquisition, and language-conditioned RL, while dramatically reducing the computational resources required to gather a large amount of experience. We compare NLE and its task suite to existing alternatives, and discuss why it is an ideal medium for testing the robustness and systematic generalization of RL agents. We demonstrate empirical success for early stages of the game using a distributed Deep RL baseline and Random Network Distillation exploration, alongside qualitative analysis of various agents trained in the environment.