greenville
ExpressEdit: Video Editing with Natural Language and Sketching
Tilekbay, Bekzat, Yang, Saelyne, Lewkowicz, Michal, Suryapranata, Alex, Kim, Juho
Informational videos serve as a crucial source for explaining conceptual and procedural knowledge to novices and experts alike. When producing informational videos, editors edit videos by overlaying text/images or trimming footage to enhance the video quality and make it more engaging. However, video editing can be difficult and time-consuming, especially for novice video editors who often struggle with expressing and implementing their editing ideas. To address this challenge, we first explored how multimodality$-$natural language (NL) and sketching, which are natural modalities humans use for expression$-$can be utilized to support video editors in expressing video editing ideas. We gathered 176 multimodal expressions of editing commands from 10 video editors, which revealed the patterns of use of NL and sketching in describing edit intents. Based on the findings, we present ExpressEdit, a system that enables editing videos via NL text and sketching on the video frame. Powered by LLM and vision models, the system interprets (1) temporal, (2) spatial, and (3) operational references in an NL command and spatial references from sketching. The system implements the interpreted edits, which then the user can iterate on. An observational study (N=10) showed that ExpressEdit enhanced the ability of novice video editors to express and implement their edit ideas. The system allowed participants to perform edits more efficiently and generate more ideas by generating edits based on user's multimodal edit commands and supporting iterations on the editing commands. This work offers insights into the design of future multimodal interfaces and AI-based pipelines for video editing.
- North America > United States > New York > New York County > New York City (0.06)
- North America > United States > South Carolina > Greenville County > Greenville (0.05)
- Asia > South Korea > Daejeon > Daejeon (0.04)
- (4 more...)
- Education > Educational Technology (0.68)
- Education > Educational Setting > Online (0.46)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.93)
Creating an African American-Sounding TTS: Guidelines, Technical Challenges,and Surprising Evaluations
Pinhanez, Claudio, Fernandez, Raul, Grave, Marcelo, Nogima, Julio, Hoory, Ron
This poses challenges for applications interested in targeting specific demographics (e.g., an African American business or NGO; a voice-tutoring system for children that are not of White ethnicity, etc.). The ultimate goal of the project described in this paper is to provide to designers, developers, and enterprises the choice of having a professional voice which is clearly recognizable as African American, and therefore more able to address diversity and inclusiveness issues. Being more precise, our goal is to create an African American Text-to-Speech system, which we will refer simply as an African American voice or AA voice, able to produce synthetic audio segments from standard English texts, and which will be recognized by African American speakers and non-speakers as sounding like a native African American speaker. The AA voice should exhibit a level of technical quality similar to the Standard American English (SAE) synthetic voices currently available through professional platforms. The evaluation of the technical quality of the AA voice, however, is not addressed in this paper, which focuses primarily on whether the AA voice can be recognized as sounding like an African American speaker. Linguists [27, 28] have described a continuum of dialects under what is often termed African American Vernacular English (AAVE). At one end of the spectrum, one finds the largest deviation from SAE in terms of lexicon (including slang), syntax and morphology, and phonological/phonetic properties. At the other end, AAVE speakers begin to approach SAE in terms of lexicon and grammar but still retain marked speech characteristics (primarily in terms of intonation, phonation, and vowel placement [14, 28]) which grant the speech a distinctive identity which listeners use as cues in the perception of African American English [44].
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > South Carolina > Greenville County > Greenville (0.06)
- North America > United States > New York > New York County > New York City (0.04)
- (31 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Questionnaire & Opinion Survey (1.00)
FARPLS: A Feature-Augmented Robot Trajectory Preference Labeling System to Assist Human Labelers' Preference Elicitation
Lyu, Hanfang, Bai, Yuanchen, Liang, Xin, Das, Ujaan, Shi, Chuhan, Gong, Leiliang, Li, Yingchi, Sun, Mingfei, Ge, Ming, Ma, Xiaojuan
Preference-based learning aims to align robot task objectives with human values. One of the most common methods to infer human preferences is by pairwise comparisons of robot task trajectories. Traditional comparison-based preference labeling systems seldom support labelers to digest and identify critical differences between complex trajectories recorded in videos. Our formative study (N = 12) suggests that individuals may overlook non-salient task features and establish biased preference criteria during their preference elicitation process because of partial observations. In addition, they may experience mental fatigue when given many pairs to compare, causing their label quality to deteriorate. To mitigate these issues, we propose FARPLS, a Feature-Augmented Robot trajectory Preference Labeling System. FARPLS highlights potential outliers in a wide variety of task features that matter to humans and extracts the corresponding video keyframes for easy review and comparison. It also dynamically adjusts the labeling order according to users' familiarities, difficulties of the trajectory pair, and level of disagreements. At the same time, the system monitors labelers' consistency and provides feedback on labeling progress to keep labelers engaged. A between-subjects study (N = 42, 105 pairs of robot pick-and-place trajectories per person) shows that FARPLS can help users establish preference criteria more easily and notice more relevant details in the presented trajectories than the conventional interface. FARPLS also improves labeling consistency and engagement, mitigating challenges in preference elicitation without raising cognitive loads significantly
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > South Carolina > Greenville County > Greenville (0.06)
- (22 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Questionnaire & Opinion Survey (1.00)
- Personal > Interview (0.67)
- Leisure & Entertainment (0.67)
- Health & Medicine (0.46)
iScore: Visual Analytics for Interpreting How Language Models Automatically Score Summaries
Coscia, Adam, Holmes, Langdon, Morris, Wesley, Choi, Joon Suh, Crossley, Scott, Endert, Alex
The recent explosion in popularity of large language models (LLMs) has inspired learning engineers to incorporate them into adaptive educational tools that automatically score summary writing. Understanding and evaluating LLMs is vital before deploying them in critical learning environments, yet their unprecedented size and expanding number of parameters inhibits transparency and impedes trust when they underperform. Through a collaborative user-centered design process with several learning engineers building and deploying summary scoring LLMs, we characterized fundamental design challenges and goals around interpreting their models, including aggregating large text inputs, tracking score provenance, and scaling LLM interpretability methods. To address their concerns, we developed iScore, an interactive visual analytics tool for learning engineers to upload, score, and compare multiple summaries simultaneously. Tightly integrated views allow users to iteratively revise the language in summaries, track changes in the resulting LLM scores, and visualize model weights at multiple levels of abstraction. To validate our approach, we deployed iScore with three learning engineers over the course of a month. We present a case study where interacting with iScore led a learning engineer to improve their LLM's score accuracy by three percentage points. Finally, we conducted qualitative interviews with the learning engineers that revealed how iScore enabled them to understand, evaluate, and build trust in their LLMs during deployment.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- (12 more...)
- Instructional Material (0.93)
- Research Report > New Finding (0.46)
- Health & Medicine (1.00)
- Education > Educational Setting > Online (1.00)
- Education > Assessment & Standards (1.00)
- Education > Educational Technology > Educational Software > Computer Based Training (0.46)
East Carolina's Parker Byrd becomes first NCAA D1 baseball athlete to play with prosthetic leg: 'It's unreal'
Fox News Flash top sports headlines are here. Check out what's clicking on Foxnews.com. Parker Byrd cemented his name in college baseball history books on Friday. The East Carolina University infielder and pitcher entered the game as a pinch-hitter. Byrd received a standing ovation from the crowd inside the stadium, who wanted to recognize him for becoming the first NCAA Division I baseball player to compete in a game with a prosthetic leg.
- North America > United States > Oklahoma (0.18)
- North America > United States > North Carolina > Pitt County > Greenville (0.07)
- North America > United States > Virginia (0.06)
- North America > United States > South Carolina (0.06)
- Leisure & Entertainment > Sports > Baseball (1.00)
- Health & Medicine (1.00)
LAVE: LLM-Powered Agent Assistance and Language Augmentation for Video Editing
Wang, Bryan, Li, Yuliang, Lv, Zhaoyang, Xia, Haijun, Xu, Yan, Sodhi, Raj
Video creation has become increasingly popular, yet the expertise and effort required for editing often pose barriers to beginners. In this paper, we explore the integration of large language models (LLMs) into the video editing workflow to reduce these barriers. Our design vision is embodied in LAVE, a novel system that provides LLM-powered agent assistance and language-augmented editing features. LAVE automatically generates language descriptions for the user's footage, serving as the foundation for enabling the LLM to process videos and assist in editing tasks. When the user provides editing objectives, the agent plans and executes relevant actions to fulfill them. Moreover, LAVE allows users to edit videos through either the agent or direct UI manipulation, providing flexibility and enabling manual refinement of agent actions. Our user study, which included eight participants ranging from novices to proficient editors, demonstrated LAVE's effectiveness. The results also shed light on user perceptions of the proposed LLM-assisted editing paradigm and its impact on users' creativity and sense of co-creation. Based on these findings, we propose design implications to inform the future development of agent-assisted content editing.
- North America > Canada > Ontario > Toronto (0.14)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > South Carolina > Greenville County > Greenville (0.05)
- (14 more...)
- Research Report (1.00)
- Questionnaire & Opinion Survey (1.00)
- Overview (0.92)
- Leisure & Entertainment (1.00)
- Health & Medicine (0.68)
Looking for a better fit? An Incremental Learning Multimodal Object Referencing Framework adapting to Individual Drivers
Gomaa, Amr, Reyes, Guillermo, Feld, Michael, Krüger, Antonio
The rapid advancement of the automotive industry towards automated and semi-automated vehicles has rendered traditional methods of vehicle interaction, such as touch-based and voice command systems, inadequate for a widening range of non-driving related tasks, such as referencing objects outside of the vehicle. Consequently, research has shifted toward gestural input (e.g., hand, gaze, and head pose gestures) as a more suitable mode of interaction during driving. However, due to the dynamic nature of driving and individual variation, there are significant differences in drivers' gestural input performance. While, in theory, this inherent variability could be moderated by substantial data-driven machine learning models, prevalent methodologies lean towards constrained, single-instance trained models for object referencing. These models show a limited capacity to continuously adapt to the divergent behaviors of individual drivers and the variety of driving scenarios. To address this, we propose \textit{IcRegress}, a novel regression-based incremental learning approach that adapts to changing behavior and the unique characteristics of drivers engaged in the dual task of driving and referencing objects. We suggest a more personalized and adaptable solution for multimodal gestural interfaces, employing continuous lifelong learning to enhance driver experience, safety, and convenience. Our approach was evaluated using an outside-the-vehicle object referencing use case, highlighting the superiority of the incremental learning models adapted over a single trained model across various driver traits such as handedness, driving experience, and numerous driving conditions. Finally, to facilitate reproducibility, ease deployment, and promote further research, we offer our approach as an open-source framework at \url{https://github.com/amrgomaaelhady/IcRegress}.
- North America > United States > South Carolina > Greenville County > Greenville (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Germany > Saarland > Saarbrücken (0.04)
- (3 more...)
- Automobiles & Trucks (1.00)
- Education > Educational Setting > Continuing Education (0.34)
Janice: Excited for eclipse
I was 8-years-old and remember being both terrified and intrigued about something that was being talked about everywhere. This wasn't a storyline out of a science fiction movie or novel, this was real, and happening here on Earth. Millions of people were going to witness something that maybe happens a couple of times in our lifetime: A total solar eclipse. Our teachers were planning lessons about this incredible celestial event. Chalkboard diagrams, planetary mobiles and handmade viewing devices were being created out of shoe boxes.
- North America > United States > Missouri > Jackson County > Kansas City (0.15)
- North America > United States > South Carolina > Greenville County > Greenville (0.06)
- North America > United States > Wyoming > Natrona County > Casper (0.05)
- (6 more...)
Learning to Prosper in a Factory Town
In the foothills of the Appalachian Mountains in a corner of South Carolina sits a town that should be economically dead. For decades, Greenville was the heart of the state's textile industry--and its economic engine. First attracted by the area's fast-moving rivers as a way to power looms, textile manufacturers employed tens of thousands of people here. Beginning in the 1970s, however, facing competition from lower-cost manufacturing regions like Mexico and Southeast Asia, these companies began to struggle. Over the next decades, many factories closed.
- North America > Mexico (0.24)
- Asia > Southeast Asia (0.24)
- North America > United States > South Carolina > Greenville County > Greer (0.04)
- (4 more...)
- Banking & Finance (1.00)
- Textiles, Apparel & Luxury Goods (0.69)
- Government > Regional Government (0.47)