Encoding of acoustic and linguistic information during language learning (mTRFxL2)

Learning a second language (L2) in adulthood is not easy and not all learners achieve the desired level of proficiency. One challenge during learning is that individuals must learn to detect acoustic and linguistic structures that differ from those of their native language. Can immersion experience help learners encode acoustic cues and linguistic structures better? Does the robustness of acoustic and linguistic encoding drive language proficiency? The primary aim of this project is to investigate how learners with different levels of proficiency and immersion experience encode acoustic and linguistic information in L2 speech. The results from this study will advance our understanding of how learners process different types of information during learning. This work will help pinpoint which aspects of language individuals are struggling to follow and which should be addressed with remedial training. This project is funded by the British Academy/Leverhulme Small Research Grants Programme.

Collaborators: Prof Adam Tierney and Dr Giovanni Di Liberto.

← Back to Research

Representations of natural sound categories (EnviSounds)

By necessity, the similarity between sounds is defined with respect to the dimensions used in making similarity judgments. Sounds might be similar according to one dimension (e.g., pitch) but dissimilar according to another dimension (e.g., amplitude). However, it is difficult to disentangle their contributions, as various properties often concur; for example, sounds with similar pitch might also have similar amplitude. Acoustic differences can also be overwritten entirely by the sounds’ meaning or common context. Thus, it is vital to assess the degree to which different object properties determine perceived similarity. This project explores whether and, if so, to what degree acoustic vs semantic information about environmental sounds contributes to similarity judgments. Building on our previous work measuring similarity between sounds, we want to extend this framework by investigating the representations of natural sound categories and various dimensions underlying those representations. We will test whether and to what degree similarity judgements depend on the dimension judged. In the real world, many dimensions are jointly and to varying degrees driving our perception of the surrounding world. Comparing the yielded representational similarity matrices (RDMs) would allow us to discover to what degree different sound properties both uniquely and in common determine perceived similarity.

Collaborators: Prof Fred Dick and Dr Jasper van den Bosch.

← Back to Research

Automated speech assessments (ACAT)

To better support language teachers and learners, I have been exploring how automated assessment tools can increase the frequency and accessibility of feedback during L2 learning. Timely feedback is critical in second language learning, yet it is often limited in both classroom and self-study contexts. In response to this challenge, my work has of late focused on the development of automated comprehensibility assessments. In one such study (Saito, Macmillan, Kachlicka, Kunihara, & Minematsu, 2022), our team uses a machine learning approach involving DNN-HMM models (Deep Neural Networks and Hidden Markov Models) trained on automated fluency measures (articulation rate and pause ratio), pitch and amplitude variation, and phonetic posteriorgrams to predict comprehensibility ratings. The resulting scores strongly correlate with human assessments. Currently, we are developing a more user-friendly tool that doesn’t require advanced modelling to obtain such ratings. Instead, it employs regression models trained on automated segmental, prosodic, and fluency measures and incorporates this analytical pipeline into an app. Our tool predicts learner comprehensibility with remarkable accuracy that generalizes across different task conditions. These projects highlight both the importance of timely, individualized feedback in L2 learning, which is scarce, and the potential of automated tools to provide effective, scalable support.

Collaborators: Prof Kazuya Saito (PI).

← Back to Research

Speech cue weighting and salience (SpeechCues)

Language structure is conveyed in speech by a complex set of acoustic cues, including changes in duration, amplitude, and frequency. Individuals who are better able to detect these acoustic cues may be able to more rapidly absorb the structure of a new language. However, detecting the acoustic cues of a second language can be difficult due to differences across languages in how sound patterns convey language structure. In particular, second language speakers may have difficulty directing attention towards the most relevant cues. Individuals who are better able to direct their attention to individual acoustic dimensions may be better able to focus on the acoustic cues which provide the most reliable information in an L2. We are testing the role of attention in second language learning using electroencephalography (EEG) and whether second language learners can be trained to attend to the most informative cues, and if so, whether this experience changes how they attend to sound cues. This project will lead to a better understanding of why some people struggle to learn a second language more than others.

Supervisors: Prof Adam Tierney (PI), Prof Fred Dick, and Prof Kazuya Saito.

← Back to Research

Last modified 2025/10/24

HTML & CSS by MLPdesign & MKachlicka