Goto

Collaborating Authors

 Educational Setting


Agentic Large Language Models, a survey

arXiv.org Artificial Intelligence

There is great interest in agentic LLMs, large language models that act as agents. We review the growing body of work in this area and provide a research agenda. Agentic LLMs are LLMs that (1) reason, (2) act, and (3) interact. We organize the literature according to these three categories. The research in the first category focuses on reasoning, reflection, and retrieval, aiming to improve decision making; the second category focuses on action models, robots, and tools, aiming for agents that act as useful assistants; the third category focuses on multi-agent systems, aiming for collaborative task solving and simulating interaction to study emergent social behavior. We find that works mutually benefit from results in other categories: retrieval enables tool use, reflection improves multi-agent collaboration, and reasoning benefits all categories. We discuss applications of agentic LLMs and provide an agenda for further research. Important applications are in medical diagnosis, logistics and financial market analysis. Meanwhile, self-reflective agents playing roles and interacting with one another augment the process of scientific research itself. Further, agentic LLMs may provide a solution for the problem of LLMs running out of training data: inference-time behavior generates new training states, such that LLMs can keep learning without needing ever larger datasets. We note that there is risk associated with LLM assistants taking action in the real world, while agentic LLMs are also likely to benefit society.


Claude's new Learning mode will prompt students to answer questions on their own

Engadget

According to a recent Digital Education Council survey, as many as 86 percent of university students globally use artificial intelligence to assist with their coursework. It's a staggering statistic that's likely to have far-reaching consequences for years to come. So it's not surprising to see a company like Anthropic announce Claude for Education, an initiative it says will equip universities to "play a key role in actively shaping AI's role in society." At the heart of Claude for Education is a new Learning mode that changes how Anthropic's chatbot interacts with users. With the feature engaged, Claude will attempt to guide students to a solution, rather than providing an answer outright, when asked a question.


Forthcoming machine learning and AI seminars: April 2025 edition

AIHub

This post contains a list of the AI-related seminars that are scheduled to take place between 1 April and 31 May 2025. All events detailed here are free and open for anyone to attend virtually. Lie-Poisson Neural Networks (LPNets): Data-Based Computing of Hamiltonian Systems Speaker: Vakhtang Poutkaradze (University of Alberta) Organised by: University of Minnesota Zoom registration is here. Sample complexity of data-driven tuning of model hyperparameters in neural networks with structured parameter-dependent dual function. Speaker: Anh Nguyen (Carnegie Mellon University) Organised by: Carnegie Mellon University Zoom link is here.


7 leadership lessons for navigating the AI turbulence

ZDNet

In a recent episode of our weekly podcast DisrupTV, Constellation Research CEO Ray Wang and I assembled an extraordinary panel of leaders to discuss effective leadership in today's rapidly changing world. The conversation featured Ellen McCarthy, founder and CEO of the Trust in Media Cooperative; Lev Gonick, award-winning CIO of Arizona State University (ASU); and Dr. David Bray, Chair of the Accelerator and Distinguished Fellow at the Stimson Center. The discussion revealed critical insights for CEOs, boards, and C-suite executives navigating today's complex leadership landscape. Here are the key takeaways from these seasoned leaders. ASU's Lev Gonick shared how the school has consistently turned moments of disruption into strategic advantages.


How and why parents and teachers are introducing young children to AI

The Guardian

Since the release of ChatGPT in late 2022, generative artificial intelligence has trickled down from adults in their offices to university students in campus libraries to teenagers in high school hallways. Now it's reaching the youngest among us, and parents and teachers are grappling with the most responsible way to introduce their under-13s to a new technology that may fundamentally reshape the future. Though the terms of service for ChatGPT, Google's Gemini and other AI models specify that the tools are only meant for those over 13, parents and teachers are taking the matter of AI education into their own hands. Inspired by a story we published on parents who are teaching their children to use AI to set them up for success in school and at work, we asked Guardian readers how and why – or why not – others are doing the same. Though our original story only concerned parents, we have also included teachers in the responses published below, as preparing children for future studies and jobs is one of educators' responsibilities as well.


Everyone's using ChatGPT, but most are doing it completely wrong

Popular Science

AI should be saving you time, boosting your productivity, and even helping you think more creatively. But if you're stuck rewriting prompts, dealing with bad responses, or wondering why it feels so basic, here's a hard truth: it's not ChatGPT … it's you. But getting your skills up to snuff is simple if you enroll in our best-selling e-degree program. It doesn't matter if you're a complete beginner, an aspiring master, or somewhere in between, you'll learn how to use ChatGPT like an expert for just 19.97 (reg. Don't worry about fitting time into your schedule--these courses are completely self-paced.


Student-Powered Digital Scholarship CoLab Project in the HKUST Library: Develop a Chinese Named-Entity Recognition (NER) Tool within One Semester from the Ground Up

arXiv.org Artificial Intelligence

Starting in February 2024, the HKUST Library further extended the scope of AI literacy to AI utilization, which focuses on fostering student involvement in utilizing state-of-the-art technologies in the projects that initiated by the Library, named "Digital Scholarship (DS) CoLab". A key focus of the DS CoLab scheme has been on cultivating talents and enabling students to utilize advanced technologies in practical context. It aims to reinforce the library's role as a catalyst and hub for fostering multidisciplinary collaboration and cultivate the "can do spirit" among university members. The Library offers 1-2 projects per year for students to engage with advanced technologies in practical contexts while supporting the Library in tackling challenges and streamlining operational tasks. The tool that introduced in this paper was mainly developed by two of the authors, Sherry Yip Sau Lai and Berry Han Liuruo, as part-time student helpers under one of our DS CoLab scheme in the 2024 Spring Semester (February to May 2024). This paper details the complete journey from ideation to implementation of developing a Chinese Named-Entity Recognition (NER) Tool from the group up within one semester, from the initial research and planning stages to execution and come up a viable product. The collaborative spirit fostered by this project, with students playing a central role, exemplifies the power and potential of innovative educational models that prioritize hands-on learning with student involvement.


The realization of tones in spontaneous spoken Taiwan Mandarin: a corpus-based survey and theory-driven computational modeling

arXiv.org Artificial Intelligence

A growing body of literature has demonstrated that semantics can co-determine fine phonetic detail. However, the complex interplay between phonetic realization and semantics remains understudied, particularly in pitch realization. The current study investigates the tonal realization of Mandarin disyllabic words with all 20 possible combinations of two tones, as found in a corpus of Taiwan Mandarin spontaneous speech. We made use of Generalized Additive Mixed Models (GAMs) to model f0 contours as a function of a series of predictors, including gender, tonal context, tone pattern, speech rate, word position, bigram probability, speaker and word. In the GAM analysis, word and sense emerged as crucial predictors of f0 contours, with effect sizes that exceed those of tone pattern. For each word token in our dataset, we then obtained a contextualized embedding by applying the GPT-2 large language model to the context of that token in the corpus. We show that the pitch contours of word tokens can be predicted to a considerable extent from these contextualized embeddings, which approximate token-specific meanings in contexts of use. The results of our corpus study show that meaning in context and phonetic realization are far more entangled than standard linguistic theory predicts.


When Autonomy Breaks: The Hidden Existential Risk of AI

arXiv.org Artificial Intelligence

AI risks are typically framed around physical threats to humanity, a loss of control or an accidental error causing humanity's extinction. However, I argue in line with the gradual disempowerment thesis, that there is an underappreciated risk in the slow and irrevocable decline of human autonomy. As AI starts to outcompete humans in various areas of life, a tipping point will be reached where it no longer makes sense to rely on human decision-making, creativity, social care or even leadership. What may follow is a process of gradual de-skilling, where we lose skills that we currently take for granted. Traditionally, it is argued that AI will gain human skills over time, and that these skills are innate and immutable in humans. By contrast, I argue that humans may lose such skills as critical thinking, decision-making and even social care in an AGI world. The biggest threat to humanity is therefore not that machines will become more like humans, but that humans will become more like machines.


Neuroplasticity in Artificial Intelligence -- An Overview and Inspirations on Drop In & Out Learning

arXiv.org Artificial Intelligence

Artificial Intelligence (AI) has achieved new levels of performance and spread in public usage with the rise of deep neural networks (DNNs). Initially inspired by human neurons and their connections, NNs have become the foundation of AI models for many advanced architectures. However, some of the most integral processes in the human brain, particularly neurogenesis and neuroplasticity in addition to the more spread neuroapoptosis have largely been ignored in DNN architecture design. Instead, contemporary AI development predominantly focuses on constructing advanced frameworks, such as large language models, which retain a static structure of neural connections during training and inference. In this light, we explore how neurogenesis, neuroapoptosis, and neuroplasticity can inspire future AI advances. Specifically, we examine analogous activities in artificial NNs, introducing the concepts of ``dropin'' for neurogenesis and revisiting ``dropout'' and structural pruning for neuroapoptosis. We additionally suggest neuroplasticity combining the two for future large NNs in ``life-long learning'' settings following the biological inspiration. We conclude by advocating for greater research efforts in this interdisciplinary domain and identifying promising directions for future exploration.