Goto

Collaborating Authors

 glue



This is the way: designing and compiling LEPISZCZE, a comprehensive NLP benchmark for Polish

Neural Information Processing Systems

The availability of compute and data to train larger and larger language models increases the demand for robust methods of benchmarking the true progress of LM training. Recent years witnessed significant progress in standardized benchmarking for English. Benchmarks such as GLUE, SuperGLUE, or KILT have become a de facto standard tools to compare large language models. Following the trend to replicate GLUE for other languages, the KLEJ benchmark\ (klej is the word for glue in Polish) has been released for Polish. In this paper, we evaluate the progress in benchmarking for low-resourced languages.





GLUE: Global-Local Unified Encoding for Imitation Learning via Key-Patch Tracking

Chen, Ye, Zhou, Zichen, Dou, Jianyu, Cui, Te, Yang, Yi, Yue, Yufeng

arXiv.org Artificial Intelligence

In recent years, visual representation learning has gained widespread attention in robotic imitation learning. However, in complex Out-of-Distribution(OOD) settings characterized by clutter and occlusion, the attention of global visual representations can be diluted or interfered, leading to degraded policy performance. The invariance of local representations for task-relevant objects offers a solution. By efficiently utilizing these local representations, training and testing data can be mapped to a more similar feature space, thereby mitigating the covariate shift problem. Accordingly, we propose GLUE, a global-local unified encoding framework for imitation learning based on key-patch tracking. GLUE selects and tracks key-patches as critical local representations by employing a text-guided mechanism. It features a novel fusion framework where global patch features query local patches to distill essential information, yielding fine-grained local features with low heterogeneity relative to the global context. This fused representation steers the robot's visual attention toward task-relevant objects and preserves precise global context, which together align the training and testing distributions into a similar and task-informative feature space, ultimately enhancing the robustness of the imitation learning policy. Experiments demonstrate that GLUE achieves strong performance across diverse tasks in both simulation and real-world settings, outperforming the strongest baseline by 17.6% in simulation, 36.3% in real-world environments, and 58.3% on real-world generalization settings. The project website of GLUE is available at https://GLUE666.github.io/.


RMT-KD: Random Matrix Theoretic Causal Knowledge Distillation

Ettori, Davide, Darabi, Nastaran, Senthilkumar, Sureshkumar, Trivedi, Amit Ranjan

arXiv.org Artificial Intelligence

Large deep learning models such as BERT and ResNet achieve state-of-the-art performance but are costly to deploy at the edge due to their size and compute demands. We present RMT-KD, a compression method that leverages Random Matrix Theory (RMT) for knowledge distillation to iteratively reduce network size. Instead of pruning or heuristic rank selection, RMT-KD preserves only informative directions identified via the spectral properties of hidden representations. RMT-based causal reduction is applied layer by layer with self-distillation to maintain stability and accuracy. On GLUE, AG News, and CIFAR-10, RMT-KD achieves up to 80% parameter reduction with only 2% accuracy loss, delivering 2.8x faster inference and nearly halved power consumption. These results establish RMT-KD as a mathematically grounded approach to network distillation.



This is the way: designing and compiling LEPISZCZE, a comprehensive NLP benchmark for Polish

Neural Information Processing Systems

The availability of compute and data to train larger and larger language models increases the demand for robust methods of benchmarking the true progress of LM training. Recent years witnessed significant progress in standardized benchmarking for English. Benchmarks such as GLUE, SuperGLUE, or KILT have become a de facto standard tools to compare large language models. Following the trend to replicate GLUE for other languages, the KLEJ benchmark\ (klej is the word for glue in Polish) has been released for Polish. In this paper, we evaluate the progress in benchmarking for low-resourced languages.


OpenAI's New Ad Shows 'Reasoning' AI Making Basic Errors

TIME - Tech

OpenAI released its most advanced AI model yet, called o1, for paying users on Thursday. The launch kicked off the company's "12 Days of OpenAI" event--a dozen consecutive releases to celebrate the holiday season. OpenAI has touted o1's "complex reasoning" capabilities, and announced on Thursday that unlimited access to the model would cost 200 per month. In the video the company released to show the model's strengths, a user uploads a picture of a wooden birdhouse and asks the model for advice on how to build a similar one. The model "thinks" for a short period and then spits out what on the surface appears to be a comprehensive set of instructions. Close examination reveals the instructions to be almost useless.