soderstrom
Age-Dependent Analysis and Stochastic Generation of Child-Directed Speech
Räsänen, Okko, Kocharov, Daniil
Child-directed speech (CDS) is a particular type of speech that adults use when addressing young children. Its properties also change as a function of extralinguistic factors, such as age of the child being addressed. Access to large amounts of representative and varied CDS would be useful for child language research, as this would enable controlled computational modeling experiments of infant language acquisition with realistic input in terms of quality and quantity. In this study, we describe an approach to model age-dependent linguistic properties of CDS using a language model (LM) trained on CDS transcripts and ages of the recipient children, as obtained from North American English corpora of the CHILDES database. The created LM can then be used to stochastically generate synthetic CDS transcripts in an age-appropriate manner, thereby scaling beyond the original datasets in size. We compare characteristics of the generated CDS against the real speech addressed at children of different ages, showing that the LM manages to capture age-dependent changes in CDS, except for a slight difference in the effective vocabulary size. As a side product, we also provide a systematic characterization of age-dependent linguistic properties of CDS in CHILDES, illustrating how all measured aspects of the CDS change with children's age.
Analysing the Impact of Audio Quality on the Use of Naturalistic Long-Form Recordings for Infant-Directed Speech Research
Blandón, María Andrea Cruz, Cristia, Alejandrina, Räsänen, Okko
Modelling of early language acquisition aims to understand how infants bootstrap their language skills. The modelling encompasses properties of the input data used for training the models, the cognitive hypotheses and their algorithmic implementations being tested, and the evaluation methodologies to compare models to human data. Recent developments have enabled the use of more naturalistic training data for computational models. This also motivates development of more naturalistic tests of model behaviour. A crucial step towards such an aim is to develop representative speech datasets consisting of speech heard by infants in their natural environments. However, a major drawback of such recordings is that they are typically noisy, and it is currently unclear how the sound quality could affect analyses and modelling experiments conducted on such data. In this paper, we explore this aspect for the case of infant-directed speech (IDS) and adult-directed speech (ADS) analysis. First, we manually and automatically annotated audio quality of utterances extracted from two corpora of child-centred long-form recordings (in English and French). We then compared acoustic features of IDS and ADS in an in-lab dataset and across different audio quality subsets of naturalistic data. Finally, we assessed how the audio quality and recording environment may change the conclusions of a modelling analysis using a recent self-supervised learning model. Our results show that the use of modest and high audio quality naturalistic speech data result in largely similar conclusions on IDS and ADS in terms of acoustic analyses and modelling experiments. We also found that an automatic sound quality assessment tool can be used to screen out useful parts of long-form recordings for a closer analysis with comparable results to that of manual quality annotation.
- Europe > Finland > Pirkanmaa > Tampere (0.04)
- North America > United States > California > Ventura County > Thousand Oaks (0.04)
- Europe > France > Île-de-France > Paris > Paris (0.04)
Here's how Amazon's Alexa AI is helping NASA become smarter at work
While you are busy giving Alexa commands to play your favourite song or book an Uber, the intelligent virtual assistant from Amazon is helping the US space agency organise daily tasks while making sense of intrinsic data-sets. According to Tom Soderstrom, IT Chief Technology and Innovation Officer at NASA's Jet Propulsion Laboratory (JPL), voice as a platform will become the next big thing once we learn to talk to digital assistants and chatbots in a fashion we do with friends and family. "If you have Alexa-controlled Amazon Echo smart speaker at home, tell her to enable the'NASA Mars' app. Once done, ask Alexa anything about the Red Planet and she will come back with all the right answers," Soderstrom said during the Amazon Web Services' (AWS) public sector summit in Washington. "This enables serverless computing where we don't need to build for scale but for real-life work cases and get the desired results in a much cheaper way. Remember that voice as a platform is poised to give 10 times faster results," Soderstrom noted on the inaugural "Earth and Space Day".
- Government > Space Agency (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)