Goto

Collaborating Authors

 hobby


Robot Talk Episode 134 – Robotics as a hobby, with Kevin McAleer

Robohub

Claire chatted to Kevin McAleer from kevsrobots about how to get started building robots at home. Kevin McAleer is a hobbyist robotics fanatic who likes to build robots, share videos about them on YouTube and teach people how to do the same. Kev has been building robots since 2019, when he got his first 3d printer and wanted to make more interesting builds. Kev has a degree in Computer Science, and because his day job is relatively hands-off, this hobby allows his creativity to have an outlet. Kev is a huge fan of Python and Micropython for embedded devices, and has a website - kevsrobots.com


I'm a committed introvert – but no AI will take away the joy I get from other people Emma Beddington

The Guardian

'I'm baffled how anyone could use AI to participate in a hobby.' 'I'm baffled how anyone could use AI to participate in a hobby.' I'm a committed introvert - but no AI will take away the joy I get from other people T his is depressing: according to the Cut, people are using AI to solve escape room puzzles and cheat at trivia nights. Surely, that is the definition of spoiling your own fun? "Like going into a corn maze and just wanting a straight line to the end," says one TikToker quoted in the article. There's also an interview with a keen reader who uses ChatGPT as a book club replacement, scraping the internet and aggregating "stimulating opinions and perspectives". All well and good (actually, no, it sounds bleak as hell) until he had a character's death spoilered in the fantasy epic he had been enjoying.


Grandfather builds the droids he was always looking for

Popular Science

Kurt Zimmerman brought Star Wars from a galaxy far, far away to Michigan. Kurt makes his droids out of wood, but they're filled and painted to look like metal. Breakthroughs, discoveries, and DIY tips sent every weekday. The wood exploded into a million pieces, covering the workshop floor. As he stood there looking at the mess he just made, Kurt Zimmerman was at a crossroads moment.


I built a desktop PC specialized for AI and I seriously regret it

PCWorld

Since AI has diffused into every aspect of the technology sector, I've been more than a little tempted to try my hand at some of AI's cooler applications. That growing temptation finally culminated in me building a desktop PC just for AI -- to try my hand at vibe coding apps just for fun. My budget wasn't that high, so for the build I landed on an AMD Ryzen 5 2400G CPU with a base clock speed of 3.6GHz, and an Nvidia RTX 3090 video card. That combination was validated by my fellow PC builders online as entirely suitable for AI, so I felt confident I was onto a good thing. My new PC worked well for my newest hobby, allowing me to dabble in making simple apps in DeepAgent.


Using Generative AI Personas Increases Collective Diversity in Human Ideation

Wan, Yun, Kalman, Yoram M

arXiv.org Artificial Intelligence

This study challenges the widely-reported tradeoff between generative AI's (GenAI) contribution to creative outcomes and decreased diversity of these outcomes. We modified the design of such a study, by Doshi and Hauser (2024), in which participants wrote short stories either aided or unaided by GenAI plot ideas[1]. In the modified study, plot ideas were generated through ten unique GenAI "personas" with diverse traits (e.g. cultural backgrounds, thinking styles, genre preferences), creating a pool of 300 story plots. While plot ideas from any individual persona showed high similarity (average cosine similarity of 0.92), ideas across different personas exhibited substantial variation (average similarity of 0.20). When human participants wrote stories based on these diverse plot ideas, their collective outputs maintained the same level of diversity as stories written without GenAI assistance, effectively eliminating the diversity reduction observed in [1]. Traditional text analytics further revealed that GenAI-assisted stories featured greater diversity in descriptive and emotional language compared to purely human-generated stories without GenAI assistance. Our findings demonstrate that introducing diversity at the AI input stage through distinct personas can preserve and potentially enhance the collective diversity of human creative outputs when collaborating with GenAI.


Substance over Style: Evaluating Proactive Conversational Coaching Agents

Srinivas, Vidya, Xu, Xuhai, Liu, Xin, Ayush, Kumar, Galatzer-Levy, Isaac, Patel, Shwetak, McDuff, Daniel, Althoff, Tim

arXiv.org Artificial Intelligence

While NLP research has made strides in conversational tasks, many approaches focus on single-turn responses with well-defined objectives or evaluation criteria. In contrast, coaching presents unique challenges with initially undefined goals that evolve through multi-turn interactions, subjective evaluation criteria, mixed-initiative dialogue. In this work, we describe and implement five multi-turn coaching agents that exhibit distinct conversational styles, and evaluate them through a user study, collecting first-person feedback on 155 conversations. We find that users highly value core functionality, and that stylistic components in absence of core components are viewed negatively. By comparing user feedback with third-person evaluations from health experts and an LM, we reveal significant misalignment across evaluation approaches. Our findings provide insights into design and evaluation of conversational coaching agents and contribute toward improving human-centered NLP applications.


EmpathyAgent: Can Embodied Agents Conduct Empathetic Actions?

Chen, Xinyan, Ge, Jiaxin, Dai, Hongming, Zhou, Qiang, Feng, Qiuxuan, Hu, Jingtong, Wang, Yizhou, Liu, Jiaming, Zhang, Shanghang

arXiv.org Artificial Intelligence

Empathy is fundamental to human interactions, yet it remains unclear whether embodied agents can provide human-like empathetic support. Existing works have studied agents' tasks solving and social interactions abilities, but whether agents can understand empathetic needs and conduct empathetic behaviors remains overlooked. To address this, we introduce EmpathyAgent, the first benchmark to evaluate and enhance agents' empathetic actions across diverse scenarios. EmpathyAgent contains 10,000 multimodal samples with corresponding empathetic task plans and three different challenges. To systematically evaluate the agents' empathetic actions, we propose an empathy-specific evaluation suite that evaluates the agents' empathy process. We benchmark current models and found that exhibiting empathetic actions remains a significant challenge. Meanwhile, we train Llama3-8B using EmpathyAgent and find it can potentially enhance empathetic behavior. By establishing a standard benchmark for evaluating empathetic actions, we hope to advance research in empathetic embodied agents. Our code and data are publicly available at https://github.com/xinyan-cxy/EmpathyAgent.


PhantomWiki: On-Demand Datasets for Reasoning and Retrieval Evaluation

Gong, Albert, Stankevičiūtė, Kamilė, Wan, Chao, Kabra, Anmol, Thesmar, Raphael, Lee, Johann, Klenke, Julius, Gomes, Carla P., Weinberger, Kilian Q.

arXiv.org Artificial Intelligence

High-quality benchmarks are essential for evaluating reasoning and retrieval capabilities of large language models (LLMs). However, curating datasets for this purpose is not a permanent solution as they are prone to data leakage and inflated performance results. To address these challenges, we propose PhantomWiki: a pipeline to generate unique, factually consistent document corpora with diverse question-answer pairs. Unlike prior work, PhantomWiki is neither a fixed dataset, nor is it based on any existing data. Instead, a new PhantomWiki instance is generated on demand for each evaluation. We vary the question difficulty and corpus size to disentangle reasoning and retrieval capabilities respectively, and find that PhantomWiki datasets are surprisingly challenging for frontier LLMs. Thus, we contribute a scalable and data leakage-resistant framework for disentangled evaluation of reasoning, retrieval, and tool-use abilities. Our code is available at https://github.com/kilian-group/phantom-wiki.


Extracting Social Connections from Finnish Karelian Refugee Interviews Using LLMs

Laato, Joonatan, Kanerva, Jenna, Loehr, John, Lummaa, Virpi, Ginter, Filip

arXiv.org Artificial Intelligence

We performed a zero-shot information extraction study on a historical collection of 89,339 brief Finnish-language interviews of refugee families relocated post-WWII from Finnish Eastern Karelia. Our research objective is two-fold. First, we aim to extract social organizations and hobbies from the free text of the interviews, separately for each family member. These can act as a proxy variable indicating the degree of social integration of refugees in their new environment. Second, we aim to evaluate several alternative ways to approach this task, comparing a number of generative models and a supervised learning approach, to gain a broader insight into the relative merits of these different approaches and their applicability in similar studies. We find that the best generative model (GPT-4) is roughly on par with human performance, at an F-score of 88.8%. Interestingly, the best open generative model (Llama-3-70B-Instruct) reaches almost the same performance, at 87.7% F-score, demonstrating that open models are becoming a viable alternative for some practical tasks even on non-English data. Additionally, we test a supervised learning alternative, where we fine-tune a Finnish BERT model (FinBERT) using GPT-4 generated training data. By this method, we achieved an F-score of 84.1% already with 6K interviews up to an F-score of 86.3% with 30k interviews. Such an approach would be particularly appealing in cases where the computational resources are limited, or there is a substantial mass of data to process.


User Profile with Large Language Models: Construction, Updating, and Benchmarking

Prottasha, Nusrat Jahan, Kowsher, Md, Raman, Hafijur, Anny, Israt Jahan, Bhat, Prakash, Garibay, Ivan, Garibay, Ozlem

arXiv.org Artificial Intelligence

User profile modeling plays a key role in personalized systems, as it requires building accurate profiles and updating them with new information. In this paper, we present two high-quality open-source user profile datasets: one for profile construction and another for profile updating. These datasets offer a strong basis for evaluating user profile modeling techniques in dynamic settings. We also show a methodology that uses large language models (LLMs) to tackle both profile construction and updating. Our method uses a probabilistic framework to predict user profiles from input text, allowing for precise and context-aware profile generation. Our experiments demonstrate that models like Mistral-7b and Llama2-7b perform strongly in both tasks. LLMs improve the precision and recall of the generated profiles, and high evaluation scores confirm the effectiveness of our approach.