Goto

Collaborating Authors

 duet




Automated Feedback on Student-Generated UML and ER Diagrams Using Large Language Models

Gürtl, Sebastian, Schimetta, Gloria, Kerschbaumer, David, Liut, Michael, Steinmaurer, Alexander

arXiv.org Artificial Intelligence

UML and ER diagrams are foundational in computer science education but come with challenges for learners due to the need for abstract thinking, contextual understanding, and mastery of both syntax and semantics. These complexities are difficult to address through traditional teaching methods, which often struggle to provide scalable, personalized feedback, especially in large classes. We introduce DUET (Diagrammatic UML & ER Tutor), a prototype of an LLM-based tool, which converts a reference diagram and a student-submitted diagram into a textual representation and provides structured feedback based on the differences. It uses a multi-stage LLM pipeline to compare diagrams and generate reflective feedback. Furthermore, the tool enables analytical insights for educators, aiming to foster self-directed learning and inform instructional strategies. We evaluated DUET through semi-structured interviews with six participants, including two educators and four teaching assistants. They identified strengths such as accessibility, scalability, and learning support alongside limitations, including reliability and potential misuse. Participants also suggested potential improvements, such as bulk upload functionality and interactive clarification features. DUET presents a promising direction for integrating LLMs into modeling education and offers a foundation for future classroom integration and empirical evaluation.


Dyads: Artist-Centric, AI-Generated Dance Duets

Wang, Zixuan, Zerkowski, Luis, Vidrin, Ilya, Pettee, Mariel

arXiv.org Artificial Intelligence

Existing AI-generated dance methods primarily train on motion capture data from solo dance performances, but a critical feature of dance in nearly any genre is the interaction of two or more bodies in space. Moreover, many works at the intersection of AI and dance fail to incorporate the ideas and needs of the artists themselves into their development process, yielding models that produce far more useful insights for the AI community than for the dance community. This work addresses both needs of the field by proposing an AI method to model the complex interactions between pairs of dancers and detailing how the technical methodology can be shaped by ongoing co-creation with the artistic stakeholders who curated the movement data. Our model is a probability-and-attention-based Variational Autoencoder that generates a choreographic partner conditioned on an input dance sequence. We construct a custom loss function to enhance the smoothness and coherence of the generated choreography. Our code is open-source, and we also document strategies for other interdisciplinary research teams to facilitate collaboration and strong communication between artists and technologists.


DUET: Optimizing Training Data Mixtures via Feedback from Unseen Evaluation Tasks

Chen, Zhiliang, Lau, Gregory Kang Ruey, Foo, Chuan-Sheng, Low, Bryan Kian Hsiang

arXiv.org Machine Learning

The performance of a machine learning (ML) model depends heavily on the relevance of its training data to the domain of the downstream evaluation task. However, in practice, the data involved in an unseen evaluation task is often not known to us (e.g., conversations between an LLM and a user are end-to-end encrypted). So, it is not obvious what data would be relevant for training/fine-tuning the ML model to maximize its task performance. Instead, one can only deploy the ML model in the unseen evaluation task to gather multiple rounds of coarse feedback on how well the model has performed. This paper presents a novel global-to-local algorithm called DUET that can exploit the feedback loop by interleaving a data selection method with Bayesian optimization. As a result, DUET can efficiently refine the training data mixture from a pool of data domains to maximize the model's performance on the unseen evaluation task and its convergence to the optimal data mixture can be theoretically guaranteed by analyzing its cumulative regret. Empirical evaluation on image and LLM evaluation tasks shows that DUET finds better training data mixtures than conventional baselines.


DUET: Dual Clustering Enhanced Multivariate Time Series Forecasting

Qiu, Xiangfei, Wu, Xingjian, Lin, Yan, Guo, Chenjuan, Hu, Jilin, Yang, Bin

arXiv.org Machine Learning

Multivariate time series forecasting is crucial for various applications, such as financial investment, energy management, weather forecasting, and traffic optimization. However, accurate forecasting is challenging due to two main factors. First, real-world time series often show heterogeneous temporal patterns caused by distribution shifts over time. Second, correlations among channels are complex and intertwined, making it hard to model the interactions among channels precisely and flexibly. In this study, we address these challenges by proposing a general framework called DUET, which introduces dual clustering on the temporal and channel dimensions to enhance multivariate time series forecasting. First, we design a Temporal Clustering Module (TCM) that clusters time series into fine-grained distributions to handle heterogeneous temporal patterns. For different distribution clusters, we design various pattern extractors to capture their intrinsic temporal patterns, thus modeling the heterogeneity. Second, we introduce a novel Channel-Soft-Clustering strategy and design a Channel Clustering Module (CCM), which captures the relationships among channels in the frequency domain through metric learning and applies sparsification to mitigate the adverse effects of noisy channels. Finally, DUET combines TCM and CCM to incorporate both the temporal and channel dimensions. Extensive experiments on 25 real-world datasets from 10 application domains, demonstrate the state-of-the-art performance of DUET.


Lenovo and Samsung have new Chromebooks you need to check out

PCWorld

Google invited me to New York to look at the newest Gemini AI tools going into Chromebooks. But I don't want to talk about that, because it's boring. Instead I want to talk about the new Chromebook models they showed me after: an updated version of Lenovo's mega-popular Chromebook Duet, and a crazy-sleek design from Samsung on the Chromebook Plus platform. The original Chromebook Duet was launched back in 2020, and it turned some heads. With a 10-inch tablet form factor plus a detachable keyboard and kickstand in the box, it was basically a perfect mix between Microsoft's Surface form factor and an iPad's ease of access.


Semantic Codebook Learning for Dynamic Recommendation Models

Lv, Zheqi, He, Shaoxuan, Zhan, Tianyu, Zhang, Shengyu, Zhang, Wenqiao, Chen, Jingyuan, Zhao, Zhou, Wu, Fei

arXiv.org Artificial Intelligence

Dynamic sequential recommendation (DSR) can generate model parameters based on user behavior to improve the personalization of sequential recommendation under various user preferences. However, it faces the challenges of large parameter search space and sparse and noisy user-item interactions, which reduces the applicability of the generated model parameters. The Semantic Codebook Learning for Dynamic Recommendation Models (SOLID) framework presents a significant advancement in DSR by effectively tackling these challenges. By transforming item sequences into semantic sequences and employing a dual parameter model, SOLID compresses the parameter generation search space and leverages homogeneity within the recommendation system. The introduction of the semantic metacode and semantic codebook, which stores disentangled item representations, ensures robust and accurate parameter generation. Extensive experiments demonstrates that SOLID consistently outperforms existing DSR, delivering more accurate, stable, and robust recommendations.


Order-Optimal Regret in Distributed Kernel Bandits using Uniform Sampling with Shared Randomness

Pavlovic, Nikola, Salgia, Sudeep, Zhao, Qing

arXiv.org Machine Learning

We consider distributed kernel bandits where $N$ agents aim to collaboratively maximize an unknown reward function that lies in a reproducing kernel Hilbert space. Each agent sequentially queries the function to obtain noisy observations at the query points. Agents can share information through a central server, with the objective of minimizing regret that is accumulating over time $T$ and aggregating over agents. We develop the first algorithm that achieves the optimal regret order (as defined by centralized learning) with a communication cost that is sublinear in both $N$ and $T$. The key features of the proposed algorithm are the uniform exploration at the local agents and shared randomness with the central server. Working together with the sparse approximation of the GP model, these two key components make it possible to preserve the learning rate of the centralized setting at a diminishing rate of communication.


Google's Bard AI officially becomes Gemini

PCWorld

If you asked a hundred people on the street what Google's generative artificial intelligence product is called, I'm betting that 98 of them wouldn't know, unless that street happens to be in San Jose. It's "Bard," for the record, unconsciously associating generative AI with British playwrights and D&D players who failed their sexual harassment prevention training. Now it's called "Google Gemini," after a brief period of rebranding, and it's debuting in Google's workplace products. Gemini is an umbrella term for Google's generative AI (a separate app on Android, embedded within the Google search app on iOS) and its integration into existing services like Gmail, Docs, Sheets, Slides, and Meet, where it was sometimes referred to as "Duet" before. Some advanced capabilities, like being a "personal tutor" or assistive writing in "more advanced coding scenarios," will be behind a Gemini Advanced paywall.