sonification
NASA offers dazzling new sights (and sounds) of the Andromeda galaxy
Breakthroughs, discoveries, and DIY tips sent every weekday. Even a century after Edward Hubble confirmed its existence, astronomers learn new details about the Andromeda galaxy that help us better understand our cosmic neighborhood and the wider universe. Earlier this week, NASA released its latest detailed images of the Milky Way's spiral sibling, as well an ethereal sonification of its energy wavelengths. Attaining an outside view of the Milky Way galaxy is a bit like trying to examine the entire planet from your backyard--that is to say, it's impossible from humanity's current vantage point. The next best option for astronomers is gazing at similar nearby spiral galaxies, the closest of which is Messier 31.
- Government > Space Agency (0.68)
- Government > Regional Government > North America Government > United States Government (0.68)
A Scene Representation for Online Spatial Sonification
Wu, Lan, Jin, Craig, Uttsha, Monisha Mushtary, Vidal-Calleja, Teresa
Robotic perception is emerging as a crucial technology for navigation aids, particularly benefiting individuals with visual impairments through sonification. This paper presents a novel mapping framework that accurately represents spatial geometry for sonification, transforming physical spaces into auditory experiences. By leveraging depth sensors, we convert incrementally built 3D scenes into a compact 360-degree representation based on angular and distance information, aligning with human auditory perception. Our proposed mapping framework utilises a sensor-centric structure, maintaining 2D circular or 3D cylindrical representations, and employs the VDB-GPDF for efficient online mapping. We introduce two sonification modes-circular ranging and circular ranging of objects-along with real-time user control over auditory filters. Incorporating binaural room impulse responses, our framework provides perceptually robust auditory feedback. Quantitative and qualitative evaluations demonstrate superior performance in accuracy, coverage, and timing compared to existing approaches, with effective handling of dynamic objects. The accompanying video showcases the practical application of spatial sonification in room-like environments.
Hearing the Robot's Mind: Sonification for Explicit Feedback in Human-Robot Interaction
Arreghini, Simone, Paolillo, Antonio, Abbate, Gabriele, Giusti, Alessandro
Social robots are required not only to understand human intentions but also to effectively communicate their intentions or own internal states to users. This study explores the use of sonification to provide explicit auditory feedback, enhancing mutual understanding in HRI. We introduce a novel sonification approach that conveys the robot's internal state, linked to its perception of nearby individuals and their interaction intentions. The approach is evaluated through a two-fold user study: an online video-based survey with 26 participants and live experiments with 10 participants. Results indicate that while sonification improves the robot's expressivity and communication effectiveness, the design of the auditory feedback needs refinement to enhance user experience. Participants found the auditory cues useful but described the sounds as uninteresting and unpleasant. These findings underscore the importance of carefully designed auditory feedback in developing more effective and engaging Human-Robot Interaction (HRI) systems.
- Questionnaire & Opinion Survey (1.00)
- Research Report > New Finding (0.34)
The Ballad of the Bots: Sonification Using Cognitive Metaphor to Support Immersed Teleoperation of Robot Teams
Simmons, Joe, Bremner, Paul, Mitchell, Thomas J, Bown, Alison, McIntosh, Verity
As an embodied and spatial medium, virtual reality is proving an attractive proposition for robot teleoperation in hazardous environments. This paper examines a nuclear decommissioning scenario in which a simulated team of semi-autonomous robots are used to characterise a chamber within a virtual nuclear facility. This study examines the potential utility and impact of sonification as a means of communicating salient operator data in such an environment. However, the question of what sound should be used and how it can be applied in different applications is far from resolved. This paper explores and compares two sonification design approaches. The first is inspired by the theory of cognitive metaphor to create sonifications that align with socially acquired contextual and ecological understanding of the application domain. The second adopts a computationalist approach using auditory mappings that are commonplace in the literature. The results suggest that the computationalist approach outperforms the cognitive metaphor approach in terms of predictability and mental workload. However, qualitative data analysis demonstrates that the cognitive metaphor approach resulted in sounds that were more intuitive, and were better implemented for spatialisation of data sources and data legibility when there was more than one sound source.
- Europe > United Kingdom > England > Bristol (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Massachusetts > Middlesex County > Reading (0.04)
- Europe > Germany > Berlin (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Energy > Power Industry > Utilities > Nuclear (0.54)
- Leisure & Entertainment > Games (0.50)
- Government > Military (0.34)
SOMson -- Sonification of Multidimensional Data in Kohonen Maps
Kohonen Maps, aka. Self-organizing maps (SOMs) are neural networks that visualize a high-dimensional feature space on a low-dimensional map. While SOMs are an excellent tool for data examination and exploration, they inherently cause a loss of detail. Visualizations of the underlying data do not integrate well and, therefore, fail to provide an overall picture. Consequently, we suggest SOMson, an interactive sonification of the underlying data, as a data augmentation technique. The sonification increases the amount of information provided simultaneously by the SOM. Instead of a user study, we present an interactive online example, so readers can explore SOMson themselves. Its strengths, weaknesses, and prospects are discussed.
- Europe > Germany > Hamburg (0.04)
- North America > United States > California > Alameda County > Livermore (0.04)
- Europe > United Kingdom > England > Tyne and Wear > Newcastle (0.04)
- (4 more...)
- Leisure & Entertainment (0.68)
- Media > Music (0.46)
- Health & Medicine (0.46)
Robotic Blended Sonification: Consequential Robot Sound as Creative Material for Human-Robot Interaction
Johansen, Stine S., Browning, Yanto, Brumpton, Anthony, Donovan, Jared, Rittenbruch, Markus
Current research in robotic sounds generally focuses on either masking the consequential sound produced by the robot or on sonifying data about the robot to create a synthetic robot sound. We propose to capture, modify, and utilise rather than mask the sounds that robots are already producing. In short, this approach relies on capturing a robot's sounds, processing them according to contextual information (e.g., collaborators' proximity or particular work sequences), and playing back the modified sound. Previous research indicates the usefulness of non-semantic, and even mechanical, sounds as a communication tool for conveying robotic affect and function. Adding to this, this paper presents a novel approach which makes two key contributions: (1) a technique for real-time capture and processing of consequential robot sounds, and (2) an approach to explore these sounds through direct human-robot interaction. Drawing on methodologies from design, human-robot interaction, and creative practice, the resulting 'Robotic Blended Sonification' is a concept which transforms the consequential robot sounds into a creative material that can be explored artistically and within application-based studies.
- North America > United States > New York > New York County > New York City (0.05)
- Oceania > Australia > Queensland > Brisbane (0.04)
Leveraging Pre-Trained Autoencoders for Interpretable Prototype Learning of Music Audio
Alonso-Jiménez, Pablo, Pepino, Leonardo, Batlle-Roca, Roser, Zinemanas, Pablo, Bogdanov, Dmitry, Serra, Xavier, Rocamora, Martín
We present PECMAE, an interpretable model for music audio classification based on prototype learning. Our model is based on a previous method, APNet, which jointly learns an autoencoder and a prototypical network. Instead, we propose to decouple both training processes. This enables us to leverage existing self-supervised autoencoders pre-trained on much larger data (EnCodecMAE), providing representations with better generalization. APNet allows prototypes' reconstruction to waveforms for interpretability relying on the nearest training data samples. In contrast, we explore using a diffusion decoder that allows reconstruction without such dependency. We evaluate our method on datasets for music instrument classification (Medley-Solos-DB) and genre recognition (GTZAN and a larger in-house dataset), the latter being a more challenging task not addressed with prototypical networks before. We find that the prototype-based models preserve most of the performance achieved with the autoencoder embeddings, while the sonification of prototypes benefits understanding the behavior of the classifier.
- South America > Uruguay (0.04)
- South America > Argentina (0.04)
- Europe > Spain (0.04)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
The applicability of transperceptual and deep learning approaches to the study and mimicry of complex cartilaginous tissues
Waghorne, J., Howard, C., Hu, H., Pang, J., Peveler, W. J., Harris, L., Barrera, O.
Complex soft tissues, for example the knee meniscus, play a crucial role in mobility and joint health, but when damaged are incredibly difficult to repair and replace. This is due to their highly hierarchical and porous nature which in turn leads to their unique mechanical properties. In order to design tissue substitutes, the internal architecture of the native tissue needs to be understood and replicated. Here we explore a combined audio-visual approach - so called transperceptual - to generate artificial architectures mimicking the native ones. The proposed method uses both traditional imagery, and sound generated from each image as a method of rapidly comparing and contrasting the porosity and pore size within the samples. We have trained and tested a generative adversarial network (GAN) on the 2D image stacks. The impact of the training set of images on the similarity of the artificial to the original dataset was assessed by analyzing two samples. The first consisting of n=478 pairs of audio and image files for which the images were downsampled to 64 $\times$ 64 pixels, the second one consisting of n=7640 pairs of audio and image files for which the full resolution 256 $\times$ 256 pixels is retained but each image is divided into 16 squares to maintain the limit of 64 $\times$ 64 pixels required by the GAN. We reconstruct the 2D stacks of artificially generated datasets into 3D objects and run image analysis algorithms to characterize statistically the architectural parameters - pore size, tortuosity and pore connectivity - and compare them with the original dataset. Results show that the artificially generated dataset that undergoes downsampling performs better in terms of parameter matching. Our audiovisual approach has the potential to be extended to larger data sets to explore both how similarities and differences can be audibly recognized across multiple samples.
Tokyo Kion-On: Query-Based Generative Sonification of Atmospheric Data
Amid growing environmental concerns, interactive displays of data constitute an important tool for exploring and understanding the impact of climate change on the planet's ecosystemic integrity. This paper presents Tokyo kion-on, a query-based sonification model of Tokyo's air temperature from 1876 to 2021. The system uses a recurrent neural network architecture known as LSTM with attention trained on a small dataset of Japanese melodies and conditioned upon said atmospheric data. After describing the model's implementation, a brief comparative illustration of the musical results is presented, along with a discussion on how the exposed hyper-parameters can promote active and non-linear exploration of the data.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.84)
- North America > United States (0.04)
- Media > Music (1.00)
- Leisure & Entertainment (1.00)
Researchers Composed New Protein Based on Sonification Using Deep Learning
Protein is of utmost importance in the human body. It is considered as the building blocks of life. Scientists, for a long, have been studying its properties and functionalities in order to improve proteins and design completely new proteins that perform new functions and processes. Recently, an innovation came into being when researchers in the United States and Taiwan explored how to create new proteins by using machine learning to translate protein structures into musical scores, presenting an unusual way to translate physics concepts across disparate domains, noted APL Bioengineering. A deep learning model has been employed to design de novo proteins, based on the interplay of elementary building blocks via hierarchical patterns.
- Asia > Taiwan (0.26)
- North America > United States > Massachusetts (0.06)